Next Article in Journal
REDA: A New Methodology to Validate Sensor Systems for Person Detection under Variable Environmental Conditions
Previous Article in Journal
A Convolutional Neural-Network-Based Training Model to Estimate Actual Distance of Persons in Continuous Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Long Short-Term Memory Neural Network with Transfer Learning and Ensemble Learning for Remaining Useful Life Prediction

1
National Engineering Research Center of Fiber Optic Sensing Technology and Networks, Wuhan University of Technology, Wuhan 430070, China
2
School of Mechanical and Electrical Engineering, Wuhan University of Technology, Wuhan 430070, China
3
School of Machinery and Automation, Wuhan University of Science and Technology, Wuhan 430081, China
*
Author to whom correspondence should be addressed.
Submission received: 20 June 2022 / Revised: 14 July 2022 / Accepted: 29 July 2022 / Published: 1 August 2022
(This article belongs to the Section Physical Sensors)

Abstract

:
Prediction of remaining useful life (RUL) is greatly significant for improving the safety and reliability of manufacturing equipment. However, in real industry, it is difficult for RUL prediction models trained on a small sample of faults to obtain satisfactory accuracy. To overcome this drawback, this paper presents a long short-term memory (LSTM) neural network with transfer learning and ensemble learning and combines it with an unsupervised health indicator (HI) construction method for remaining-useful-life prediction. This study consists of the following parts: (1) utilizing the characteristics of deep belief networks and self-organizing map networks to translate raw sensor data to a synthetic HI that can effectively reflect system health; and (2) introducing transfer learning and ensemble learning to provide the required degradation mechanism for the RUL prediction model based on LSTM to improve the performance of the model. The performance of the proposed method is verified by two bearing datasets collected from experimental data, and the results show that the proposed method obtains better performance than comparable methods.

1. Introduction

In mechanical equipment, rolling element bearings are one of the most critical and vulnerable components due to their transferring motion and withstanding loads during the operation of machinery [1,2]. To guarantee the normal operation of machinery and to avoid catastrophic events, condition-based maintenance (CBM) or predictive maintenance (PM) need to be arranged using remaining-useful-life (RUL) predictions of bearings [3]. Thus, the accuracy of RUL prediction is greatly significant for the safety and reliability of manufacturing equipment.
With increasing complexity of mechanical equipment, traditional RUL approaches are unable to suit the correspondingly higher standards and higher levels for mechanical equipment. Therefore, data-driven RUL methods have been widely studied in recent years [4,5]. The framework of data-driven methods mainly consists of data acquisition, construction of health indicators (HI) and RUL prediction [6]. According to a recent review, deep learning has attracted increasing attention in the field of RUL prediction. Lei et al. [7] utilized a multi-scale dense-gate recurrent neural network to identify relevant information at different timescales to improve the performance of model. Yang et al. [8] proposed an improved long-short term memory neural network to estimate bearing performance degradation. In [9], a hybrid model based on long short-term memory and Elman neural networks predicted RUL for lithium-ion batteries. Despite existing prognosis methods achieving satisfactory accuracy, there are still two deficiencies that limit their industrial extension and application.
The first shortcoming is that most HI construction methods adopt hand-designed labels for supervised training, which makes HI construction time-consuming and laborious. A deep belief network (DBN) is a deep learning method [10] that can extract the most-suitable feature representations from raw data due to its nonlinear network structure. It has been gradually applied in the field of prognostic and health management (PHM) [11]. Peng et al. [12] utilized deep belief networks with an unsupervised algorithm to construct HI and achieved RUL prediction by the improved particle filter. Xu et al. [13] proposed an improved unsupervised deep belief network to estimate bearing performance degradation.
The second shortcoming is that RUL prediction models are trained based on the availability of large training datasets with fault trends. However, in real conditions, we can only obtain relatively small sets of fault data for the predicted target. This makes it difficult to acquire high accuracy in RUL prediction. To overcome this drawback, transfer learning and ensemble learning are provided, solving the problem of insufficient accuracy due to lack of historical measurement data. Transfer learning uses the knowledge learned from the source domain to improve learning in the target domain. Ensemble learning is a learning approach in which results are combined by multiple learning algorithms to improve performance. In the field of CBM, transfer learning and ensemble learning have been applied mainly to fault diagnosis, but there are a few RUL prediction applications. For instance, Shen et al. [14] proposed deep convolutional neural networks with ensemble learning and transfer learning to estimate the capacity of lithium-ion batteries. Zhang et al. [15] presented an instance-based ensemble deep transfer learning network to recognize degradation of ball screws. In [16], transfer component analysis (TCA) was used to find the common feature representation between different bearings, and then an SVM prediction model was constructed to achieve RUL prediction. Zhang et al. [17] utilized an ensemble learning-based prognostic approach to improve the performance of models in RUL prediction of aircraft engines.
To overcome the aforementioned drawbacks, this paper presents a hybrid model of HI construction based on a deep belief network (DBN) and a self-organizing map (SOM), and an RUL prediction model based on a long short-term memory neural network (LSTM) with transfer learning and ensemble learning. As shown in Figure 1, HI is first constructed from measured vibration data by the hybrid model based on DBN and SOM. Following that, a long short-term memory neural network with transfer learning and ensemble learning is constructed to improve the performance RUL prediction of target bearing by using the auxiliary data of the inherent degradation trend under different working conditions. In the end, the performance of the proposed method is demonstrated by experimental bearing datasets. The main contributions of this paper are summarized as follows.
(1)
Unsupervised HI construction based on DBN and SOM is used to construct HI. The process of feature extraction and feature fusion does not require artificial labels. Further, there is no need to determine a fault threshold (FT) based on the experience of researchers. Therefore, the constructed HI can eliminate the influence of personnel participation.
(2)
Transfer learning and ensemble learning are introduced into a long short-term memory neural network to improve the accuracy and robustness of the model when it is trained with a small amount of data. Experimental results indicate that LSTM with transfer learning and ensemble learning (LSTM-ETL) has better performance than the LSTM.
The rest of the paper is organized as follows: Section 2 reviews existing studies related to the basic theory. In Section 3, the proposed method is introduced. In Section 4, the performance of the proposed method is validated through an experimental dataset. Finally, conclusions are drawn in Section 5.

2. Related Basic Theory

2.1. Deep Belief Network

A deep belief network is stacked by multiple restricted Boltzmann machines (RBMs); as a deep learning method, it can extract essential patterns from raw data based on its strong nonlinear mapping performance [18]. For instance, Peng et al. [12] proposed a deep belief network to construct HI and employed an improved particle filter model for RUL prediction.
In a deep belief network, the training process is divided into an unsupervised pre-training phase and a supervised fine-tuning phase. In the pre-training phase, the output of the previous RBM is invoked as the input of the next RBM to update network parameters through a contrastive divergence (CD) algorithm. In the fine-tuning phase, a back-propagation (BP) algorithm is employed to update network parameters from the bottom layer to the top layer, as shown in Figure 2.

2.2. Long Short-Term Memory Neural Network

Compared with a traditional feed-forward neural network, a recurrent neural network (RNN) is a deep learning model into which a notion of time has been introduced. It is able to relate time and data. As the time sequence increases, the RNN [19] structure deepens by hidden-layer back-propagation. RNNs can have problems of gradient explosions or gradient vanishing in the training process.
Therefore, for long-time-sequence data, RNNs cannot remember the relationship between the current information and the long-time information in the time sequence. To solve these kinds of problems, a memory cell is used in LSTM architecture. As shown in Figure 3, the memory cell replaces the hidden layer of a traditional RNN, and it contains three kinds of gate structure: the forget gate, the input gate, and the output gate. Therefore, in time sequences, LSTMs are more suitable than traditional RNN for predicting data.
The input of LSTM consists of the previous time output h t 1 and the current time input x t , which is input to the memory cells to determine the information discarded by the forget gate. The forget gate f t controls what information from the previous memory cells is to be discarded, as defined by:
f t = σ ( w fx x t + w fh h t 1 + b f )
where σ indicates the sigmoid function; w fx and w fh indicate the weight from the input layer to the hidden layer of the forget gate and the weight from the previous hidden layer to the hidden layer of the forget gate at time t, respectively; and b f is the bias of the forget gate.
Then, the memory cells need to determine the updated new information combined with the input gate i t , the candidate value g t , the forget gate f t , and the previous state c t 1 . The following formula can be used to describe the process in which the memory cells determine the updated information.
i t = σ ( w ix x t + w ih h t 1 + b i )
g t = σ ( w gx x t + w gh h t 1 + b g )
c t = f t c t 1 + i t g t
where w ix and w ih indicate the weight from the input layer to the hidden layer of the input gate and the weight of the previous hidden layer to the hidden layer of the input gate at time t, respectively; b i and b g are the bias of the input gate and the candidate gate, respectively; c t indicates the cell state at the present time; tanh(·) indicates a hyperbolic tangent function; and indicates element-wise multiplication.
Finally, in the memory cells, the network decides the output by the current state c t and the output gate o t , which is defined as:
o t = σ ( w ox x t + w oh h t 1 + b o )
h t = o t tanh ( c t )
where w ox and w oh indicate the weight from the input layer to the hidden layer of the output gate and the weight of the previous hidden layer to the hidden layer of the output gate at time t respectively; and b o is the bias of the output gate.

3. Methodology

3.1. Proposed HI Construction

It is well-known that the performance of bearings gradually decreases with time. Correspondingly, to represent the health state of a bearing, an effective and universal method of HI construction is needed. This section mainly provides the procedure for HI construction based on unsupervised learning. The key idea is that the extracted features should retain the essential information of the raw data. Since DBN has strong nonlinear mapping performance, it is able to extract essential patterns from raw data. With data collected from n time steps T , the collected data T is firstly preprocessed to [0, 1] by normalizing according to the corresponding formula t * = ( t t min ) / ( t t max ) , where t is raw data, t min is the minimum value of raw data, and t max denotes the maximum value of raw data. After the DBN structure is determined, the normalized training dataset T * is employed to train DBN. Then, the trained DBN is obtained by the pre-training phase and the fine-tuning phase using the normalized dataset. Finally, the extracted features h n are expressed as follows:
{ h 1 = σ ( W 1 t + b 1 ) h 2 = σ ( W 2 t 1 + b 2 ) . h n = σ ( W n t n 1 + b n )
where W i , b i are parameters of DBN (i = 1, 2,…, n), x denotes the raw vibration signal, σ denotes the activation function of DBN, n is the number of hidden layers, and h i denotes the output of the ith hidden layer.
To fully utilize the feature information extracted from the raw data, it is necessary to fuse the features through an appropriate method. A self-organizing map is an unsupervised algorithm that is only composed of an input layer and a competitive layer; it can map high-dimensional data into a two-dimensional topology structure. Therefore, to take advantage of degraded information from the extracted features, an SOM is employed to fuse the extracted features. After the SOM structure is determined, the extracted training feature h n is employed to train the SOM. Then, the trained SOM is obtained by a competitive formula of training features. Finally, the feature-set is feed into the trained SOM to construct HI according to the corresponding formula h i = f m B M U , where h i is the HI at time i, f denotes the large features of input, and m B M U stands for the weight vector of the best matching unit (BMU) of input f [20]. The flowchart of the proposed HI construction based on unsupervised learning is shown in Figure 4.

3.2. Model Construction Process

3.2.1. LSTM Parameter Transfer

As mentioned before, the LSTM algorithm has the inevitable problem of low prediction accuracy with small sets of fault data. Hence, we introduce learned knowledge from the source domain to improve the performance of LSTM in the target domain. The source dataset is usually different but related to the target domain. Based on the HI constructed in Section 3.1, the source dataset and the target dataset are expressed as follows:
Y s = { H I 1 s , H I 2 s , , H I n s }
Y t = { Y t t r a i n , Y t t e s t }   with   { Y t t r a i n = { H I 1 t , H I 2 t , , H I k t } Y t t e s t = { H I k t , H I k + 1 t , , H I m t }
where Y s denotes source HI data, Y t denotes target HI data, Y t t r a i n and Y t t e s t are target training HI data and target testing HI data, respectively, H I i s denotes the HI value of the source at time i, H I i t denotes the HI value of the target HI at time i, n is the length of the source HI data, k is the length of target training HI data, and m is the length of target HI data.
The source dataset Y s is first used to pre-train n individual LSTMi ( i = 1 , 2 , , 5 ) models. Subsequently, the knowledge learned from the source dataset is employed to help complete the targeted task through transfer learning. As shown in Figure 5, the learned parameters θ i of trained LSTM i models are transferred to LSTM - TL i models of the target. Finally, the training data of target Y t t r a i n are utilized to fine-tune the parameters of LSTM - TL i models to fit the targeted task.
In parameters transfer, the parameters learned from the source domain can be effectively utilized to optimize the target model. It uses the knowledge learned from the source domain to improve the learning task in the target domain. This improves the RUL prediction accuracy of LSTM models that have a small set of training data.

3.2.2. Ensemble LSTM-ETL

After parameter transfer, a single LSTM model may have the risk of poor performance. In order to avoid this problem, ensemble learning is utilized to overcome this drawback. Therefore, the main goal of ensemble learning is to decrease the risk of creating a learning algorithm with poor performance. The framework of the LSTM-ETL model is shown in Figure 6.
In previous research, a widely used optimization method with random gradient descent (SGD) and momentum has been used to minimize the expected generalization error between the output and the real value. This maximized algorithm accuracy and robustness. The cost function of the model is defined as
C R = C + λ θ ( w ) = 1 2 t = 0 T y t y ¯ t 2 2 + λ 2 w T w
where C is the cost function of the model, λ is the L2 regularization factor of the model, θ ( w ) is the regularization term that relates to the weight, y t is the output value of the model at time t, and y ¯ t is the label value of the model at time t.
The learned parameters of n LSTM models are transferred to construct n LSTM-TL models. Subsequently, the LSTM-ETL model is established by using a BP neural network to integrate n individual LSTM-TL models. Detailed steps of the ensemble strategy are described as follows:
(1)
Construction of LSTM-ETL model: The output of n LSTM-TL models is integrated by a BP neural network. This layer is used to assign the model weights w i e to y t i and to calculate the ensemble predictive output y t e n s e m b l e of n LSTM-TL models based on weights w i e such that y t e n s e m b l e = f ( i = 1 n y t i · w i e + b ) .
(2)
Training algorithm of LSTM-ETL model: To obtain the health degradation model of the system, the LSTM-TL model is used to realize system tracking, with the HI values of k (k = 5) consecutive time points x t = [ H I t k + 1 , H I t k + 2 , , H I t ] as input and the HI value of the prospective time points y t = [ H I t + 1 ] as labels. In the next iteration, the output and input of the model are x t + 1 = [ H I t k + 2 , H I t k + 3 , , H I t + 1 ] and y t + 1 = [ H I t + 2 ] , respectively. In each iteration, the input of the model consists of the second HI value to the last HI value of the input in the previous iteration as well as the prediction output from the previous iteration [21]. By the method described above, n LSTM-TL models contain a set of parameters that are adjusted. To further update these parameters and the parameters of the secondary model (BP), we use an optimization method with random gradient descent (SGD) and momentum to update the parameters (weights, w, and biases, b) to minimize the expected generalization error between the output of LSTM-ETL and the real value. By the predicted value y t and the real value y ¯ t , the parameters of the feedforward neural network can be obtained by error back-propagation of the cost function.
(3)
Prediction of LSTM-ETL model: After the degradation LSTM-ETL model has been obtained, the future HI can be achieved by introducing the prediction of the previous step into the LSTM-ETL model. In each iteration, the inputs of n LSTM-TL models are made up of the last time k − 1 input from the previous iteration and the last time-step prediction output from the previous iteration. For example, if the input of the previous time is x t 1 = [ H I t k + 1 , H I t k + 2 , , H I t ] , then the input of the current time is x t = [ H I t k + 2 , H I t k + 3 , , y t 1 e n s e m b l e ] . The prediction results of n LSTM-TL models are integrated into a final result by the BP neural network. When the final result is below the predefined threshold for the first time, the model is terminated to obtain the RUL value. Finally, the formula for calculating the final RUL result is as follows:
x t = [ H I t k + 2 , H I t k + 3 , , y t 1 e n s e m b l e ]
y ( t ) i = R e l u ( w y h i h t i + b y i )
y t e n s e m b l e = f ( i = 1 n y t i · w i e + b )
R U L p r e d i c t = { k | y ( k ) e n s e m b l e y t h r e s h o l d }
where w y h is the weight between a hidden layer and an output layer, the vector b y is the bias parameter of the output layer, y ( k ) denotes the predicted value of the model in step k, and y t h r e s h o l d denotes the predefined failure threshold.

4. Experiment Verification

4.1. An Experiment System

Experimental data are provided by the Key Laboratory of Education Ministry for Modern Design and Rotor-Bearing System [21], which has been widely employed to research RUL prediction. In the experiments, two accelerometers are positioned on the horizontal and vertical directions of the tested bearings; one is mounted on the horizontal axis and the other one is mounted on the vertical axis. The data of tested bearings were acquired on the experimental platform shown in Figure 7. Sampling frequency was set to 25.6 kHz. A total of 32,768 datapoints (i.e., 1.28 s) were recorded for each sampling, and the sampling period was equal to 1 min. To avoid redundancy, the horizontal vibration was only employed to verify the algorithms in this paper.
In the experiments, there were three different operating conditions: (1) 2100 rpm and 12 kN, (2) 2250 rpm and 11 kN, and (3) 2400 rpm and 10 kN. The full cycle data data of Condition 2 is shown in Figure 8. The last two datasets were selected in this paper to demonstrate the effectiveness of the proposed method. Bearing2_1 data and bearing2_2 data from Condition 2 were regarded as the training dataset for constructing the HI model of the target domain. Bearing3_1 data and bearing3_2 data from Condition 3 were regarded as the training dataset for constructing the HI model of the source domain.

4.2. Health Indicator Construction

According to [18], the structure of DBN is composed of two RBMs, where the number of hidden layers is set to three. The size of the input layer is 3000. The size of the first hidden layer, the second hidden layer, and the final hidden layer is 1000, 500, and 100, respectively. The input layer of SOM is 100, and the output layer is 1. After the structure of DBN and SOM are determined, the HI is constructed following the process shown in Figure 3.
In order to verify the performance of the proposed construction method, the RMS [23], PCA [24], and DBN-SOM methods are compared for bearing3_1. As shown in Figure 9, compared with the constructed HI curves of the traditional methods, the degradation curve constructed by the DBN-SOM method is smoother and has better monotonicity. To further illustrate the effectiveness of the DBN-SOM method, the correlation and monotonicity are used to evaluate the constructed HI. The former represents the linear relationship between the constructed HI and the sampling time point. The latter evaluates the increasing or decreasing trend of the constructed HI over time. The formula is as follows:
C o r r = | t = 1 T ( F t F ˇ ) ( l t l ˇ ) | t = 1 T ( F t F ˇ ) 2 ( l t l ˇ ) 2
M o n = | N u m   o f   d F > 0 T 1 N u m   o f   d F < 0 T 1 |
where F t represents the HI value of the sample at time t, l t represents the time value of the sampling point, d F represents the differential of the HI sequence, and T represents the sampling points of the whole bearing cycle.
Table 1 shows the monotonicity and correlation results of the HI curves for the RMS, PCA, and DBN-SOM methods. It can be seen that the HI performance of the DBN-SOM method on most bearings is significantly better than the two methods used for comparison. Finally, the constructed HIs are shown in Figure 10. It can be seen that the HI constructed by the DBN-SOM method can effectively reflect the state of the bearing.
The process of bearing degradation is usually divided into two stages. In the first stage, the bearing is in a healthy state with no risk of bearing failure. In the second stage, the bearing is at risk of failure. If the training data from the first stage are included to train the model, the irrelevant time-series data will interfere with the construction of the model. Therefore, partial data without the trend of degradation is ignored. In the second stage of the constructed HI value, the first 20% of the data are used for training data, and the last 80% of the data are used for validation data.

4.3. RUL Prediction

The main objective of the proposed method is to improve the RUL prediction accuracy of LSTM in a small set of training samples. To achieve this objective, five LSTM models first are pre-trained through the source HI data, and then the LSTM-ETL model is retrained by the target HI data. Finally, the model parameter settings are shown in Table 2.
To demonstrate the performance of the LSTM-ETL model for RUL prediction, the proposed method is compared with the LSTM, LSTM-TL, and SVM methods. The trained set is used to update the RUL model; the value of RUL can be calculated when the predicted HI reaches the failure threshold, which is set 1.0.
To compare the performance of the LSTM-ETL model, two methods to calculate the error of prediction results are applied to evaluate the performance of the RUL prediction, which are defined as follows:
E r i = A c t R U L i P r e R U L i A c t R U L i × 100 %
MAE = 1 n i = 1 n | E r i |
where A c t R U L i denotes the true RUL of the ith bearing data, P r e R U L i denotes the predicted RUL of the ith bearing data, E r i denotes the average error of ith bearing data, and MAE denotes the mean absolute error of all bearing data.

4.4. Results and Discussion

The results of prediction are shown in Figure 11, in which the evaluation result of the proposed method is compared with other methods for accuracy. As shown in Figure 11, in bearings2_5, the predicted RULs based on the SVM, LSTM, LSTM-TL, and LSTM-ETL models are 18,100 s, 16,600 s, 13,300 s, and 9990 s, respectively. The corresponding actual RUL is 8570 s. The error of the LSTM-ETL model is less than the error of the SVM, LSTM, and LSTM-TL models. As mentioned before, smaller error indicates more-accurate prediction. Thus, it is proved that the proposed method can improve model performance with a small set of training samples.
In order to further illustrate the performance of the proposed method, five bearings under Condition 2 are selected as the target bearings to be predicted. The LSTM-ETL, LSTM-TL, LSTM, and SVM models are employed to predict RUL. Table 3 shows the results of the RUL predicted by the LSTM, LSTM-TL, LSTM-ETL, and SVM models in terms of both the Er (percentage error) and MAE (mean absolute error). The following two significant conclusions can be drawn from the results listed in Table 3.
(1)
Based on the overall MAE, the performance of the LSTM-ETL method is better than that of the LSTM, LSTM-TL, and SVM methods. From the perspective of Er in individual bearings, the proposed method has lower error for all four bearings except bearing2_4. When there is insufficient training data, the overall performance of the LSTM-ETL method is better than that of the other methods.
(2)
LSTM-ETL shows a MAE of 31.89%, which shows that the proposed method can precisely predict the value of RUL with a small set of data. Comparing with the MAEs of LSTM (63.39%) and SVM (66.07%), it can be concluded that both transfer learning and ensemble learning are conducive to the LSTM model’s higher prognostic accuracy. The MAE of the LSTM-ETL model is 13.54% higher than that of the LSTM-TL model, which shows that ensemble learning can effectively improve the performance of RUL prediction models for single-transfer learning.

5. Conclusions

In this study, an HI based on DBN and SOM was proposed and constructed to enhance RUL prediction accuracy of bearings. During the construction of the HI, the process of feature extraction and feature fusion did not need artificial labels. Further, there was no need to determine the fault threshold (FT) based on researcher experience. Subsequently, LSTM with ensemble learning and transfer learning (LSTM-ETL) was utilized to predict the RUL of bearings. The results of the transfer learning and ensemble learning method had relatively low MAEs. The results showed that transfer learning can improve the performance of prediction models by using the knowledge learned from the source domain to improve learning in the target model, and ensemble learning can improve accuracy by integrating the results from multiple models. In addition, the proposed method performs better than the other methods, which provides an improved strategy for RUL prediction. For the negative transfer problem in transfer learning and ensemble learning, evaluation and analyzation of bearing-data information loss under different working conditions will be the topic of future work. This is a key problem that a prediction model with transfer learning and ensemble learning can be addressed through deep metric learning.

Author Contributions

Conceptualization, C.Z.; methodology, L.W. and C.Z.; software, L.W.; validation, C.Z. and D.F.; formal analysis, L.W.; investigation, L.W. and C.Z.; data curation, L.W.; writing—original draft preparation, L.W.; writing—review and editing, C.Z.; visualization, L.W.; supervision, H.L., Z.P., D.F. and Z.W.; project administration, C.Z.; funding acquisition, C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (NSFC) (grant nos. 52071245, 61975157) and the National Key Research and Development Program of China (no. 2021YFB3202901).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support this study are available at https://biaowang.tech/xjtu-sy-bearing-datasets/ (accessed on 27 January 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Y.; Jing, L.; Zuo, M.J.; He, Z. Condition monitoring and fault diagnosis of planetary gearboxes: A review. Measurement 2014, 48, 292–305. [Google Scholar] [CrossRef]
  2. Ghods, A.; Lee, H.H. Probabilistic frequency-domain discrete wavelet transform for better detection of bearing faults in induction motors. Neurocomputing 2016, 188, 206–216. [Google Scholar] [CrossRef]
  3. Jiang, J.; Lee, J.; Zeng, Y. Time Series Multiple Channel Convolutional Neural Network with Attention-Based Long Short-Term Memory for Predicting Bearing Remaining Useful Life. Sensors 2020, 20, 166. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhao, Z.; Liang, B.; Wang, X.; Lu, W. Remaining Useful Life Prediction of Aircraft Engine based on Degradation Pattern Learning. Reliab. Eng. Syst. Saf. 2017, 164, 74–83. [Google Scholar] [CrossRef]
  5. Hu, C.; Youn, B.D.; Wang, P.; Yoon, J.T. Ensemble of data-driven prognostic algorithms for robust prediction of remaining useful life. Reliab. Eng. Syst. Saf. 2012, 103, 120–135. [Google Scholar] [CrossRef] [Green Version]
  6. Cheng, Y.; Hu, K.; Wu, J.; Zhu, H.; Lee, C.K.M. A deep learning-based two-stage prognostic approach for remaining useful life of rolling bearing. Appl. Intell. 2021, 52, 5880–5895. [Google Scholar] [CrossRef]
  7. Ren, L.; Cheng, X.; Wang, X.; Cui, J.; Zhang, L. Multi-scale Dense Gate Recurrent Unit Networks for bearing remaining useful life prediction. Future Gener. Comput. Syst. 2018, 94, 601–609. [Google Scholar] [CrossRef]
  8. Yang, J.; Peng, Y.; Xie, J.; Wang, P. Remaining Useful Life Prediction Method for Bearings Based on LSTM with Uncertainty Quantification. Sensors 2022, 22, 4549. [Google Scholar] [CrossRef] [PubMed]
  9. Li, X.; Zhang, L.; Wang, Z.; Dong, P. Remaining useful life prediction for lithium-ion batteries based on a hybrid model combining the long short-term memory and Elman neural networks. J. Energy Storage 2019, 21, 510–518. [Google Scholar] [CrossRef]
  10. Hinton, G.E.; Osindero, S.; Teh, Y.W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  11. Chong, Z.; Lim, P.; Qin, A.K.; Tan, K.C. Multiobjective Deep Belief Networks Ensemble for Remaining Useful Life Estimation in Prognostics. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2306–2318. [Google Scholar]
  12. Peng, K.; Jiao, R.; Dong, J.; Pi, Y. A deep belief network based health indicator construction and remaining useful life prediction using improved particle filter. Neurocomputing 2019, 361, 19–28. [Google Scholar] [CrossRef]
  13. Xu, F.; Fang, Z.; Tang, R.; Li, X.; Tsui, K.L. An unsupervised and enhanced deep belief network for bearing performance degradation assessment. Measurement 2020, 162, 107902. [Google Scholar] [CrossRef]
  14. Huynh, B.Q.; Li, H.; Giger, M.L. Digital mammographic tumor classifcation using transfer learning from deep convolutional neural networks. J. Med. Imaging 2016, 3, 034501. [Google Scholar] [CrossRef] [PubMed]
  15. Li, S.; Fu, Y. Unsupervised transfer learning via low-rank coding for image clustering. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA; pp. 1795–1802. [Google Scholar]
  16. Kim, J.; Kim, Y.; Sarikaya, R.; Fosler-Lussier, E. Cross-lingual transfer learning for pos tagging without cross-lingual resources. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2832–2838. [Google Scholar]
  17. Zhang, L.; Guo, L.; Gao, H.; Dong, D.; Fu, G.; Hong, X. Instance-based ensemble deep transfer learning network: A new intelligent degradation recognition method and its application on ball screw. Mech. Syst. Signal Process. 2020, 140, 106681. [Google Scholar] [CrossRef]
  18. Yan, J.; Gao, Y.; Yu, Y.; Xu, H.; Xu, Z. A Prediction Model Based on Deep Belief Network and Least Squares SVR Applied to Cross-Section Water Quality. Water 2020, 12, 1929. [Google Scholar] [CrossRef]
  19. Zheng, C.; Wang, S.; Liu, Y.; Liu, C. A novel RNN based load modelling method with measurement data in active distribution system. Electr. Power Syst. Res. 2019, 166, 112–124. [Google Scholar] [CrossRef]
  20. Fischer, A.; Igel, C. Training restricted Boltzmann machines: An introduction. Pattern Recognit. 2014, 47, 25–39. [Google Scholar] [CrossRef]
  21. Zhang, J.; Wang, P.; Yan, R.; Gao, R.X. Long short-term memory for machine remaining life prediction. J. Manuf. Syst. 2018, 48, 78–86. [Google Scholar] [CrossRef]
  22. Available online: https://biaowang.tech/xjtu-sy-bearing-datasets/ (accessed on 27 January 2021).
  23. Shen, Z.; Chen, X.; He, Z.; Sun, C.; Liu, Z. Remaining life predictions of rolling bearing based on relative features and multivariable support vector machine. J. Mech. Eng. 2013, 49, 183–189. [Google Scholar] [CrossRef]
  24. Wang, T. Bearing life prediction based on vibration signals: A case study and lessons learned. In Proceedings of the Prognostics and Health Management, Denver, CO, USA, 18–21 June 2012; IEEE: Piscataway, NJ, USA; pp. 1–7. [Google Scholar]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Sensors 22 05744 g001
Figure 2. Unsupervised fine-tuning of DBN.
Figure 2. Unsupervised fine-tuning of DBN.
Sensors 22 05744 g002
Figure 3. Diagram of LSTM cell.
Figure 3. Diagram of LSTM cell.
Sensors 22 05744 g003
Figure 4. Flowchart of bearing health indicator construction.
Figure 4. Flowchart of bearing health indicator construction.
Sensors 22 05744 g004
Figure 5. The process of transfer learning.
Figure 5. The process of transfer learning.
Sensors 22 05744 g005
Figure 6. Construction of the proposed LSTM-ETL mode.
Figure 6. Construction of the proposed LSTM-ETL mode.
Sensors 22 05744 g006
Figure 7. Rolling-bearing-life data-acquisition-experiment platform [22].
Figure 7. Rolling-bearing-life data-acquisition-experiment platform [22].
Sensors 22 05744 g007
Figure 8. The full cycle data bearing vibration signal.
Figure 8. The full cycle data bearing vibration signal.
Sensors 22 05744 g008
Figure 9. HI curve of bearing3_1 for RMS method, PCA method, and DBN-SOM method.
Figure 9. HI curve of bearing3_1 for RMS method, PCA method, and DBN-SOM method.
Sensors 22 05744 g009
Figure 10. HIs of bearings.
Figure 10. HIs of bearings.
Sensors 22 05744 g010
Figure 11. Results of RUL prediction for bearings2_1, bearings2_2, bearings2_3, bearings2_4, and bearings2_5.
Figure 11. Results of RUL prediction for bearings2_1, bearings2_2, bearings2_3, bearings2_4, and bearings2_5.
Sensors 22 05744 g011aSensors 22 05744 g011b
Table 1. The monotonicity and correlation results of HI curves for RMS, PCA, and DBN-SOM methods.
Table 1. The monotonicity and correlation results of HI curves for RMS, PCA, and DBN-SOM methods.
BearingRMSPCADBN-SOM
CorMonCorMonCorMon
2_10.260.170.270.160.290.17
2_20.650.150.650.180.710.21
2_30.510.130.520.140.510.14
2_40.190.140.210.130.310.15
2_50.550.170.560.180.610.17
3_10.310.140.320.140.400.16
3_20.170.130.170.150.300.14
3_30.430.160.410.170.450.17
s3_50.270.110.280.130.300.15
Table 2. Parameter values used in LSTM pre-training and LSTM-ETL retraining.
Table 2. Parameter values used in LSTM pre-training and LSTM-ETL retraining.
ParameterPre-TrainingRetraining
Initial learning rate0.010.001
Momentum0.90.9
Number of neurons[100, 100][100, 100]
Number of epochs10001000
Table 3. Prediction results.
Table 3. Prediction results.
Testing BearingCurrent Time (s)Actually RUL (s)LSTM-ETL Predict RUL(s) LSTM-ETL Error (%)LSTM-TL Error (%)LSTM Error (%)SVM Error (%)
2_127,4201750153012.5719.3638.1834.32
2_234806990433038.0653.27−70.96−65.38
2_320,94010,010756024.4728.1668.3670.28
2_4185059019067.8071.1945.7649.15
2_510,02085709990−16.57−55.19−93.70−111.20
MAE 31.8945.4363.3966.07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, L.; Liu, H.; Pan, Z.; Fan, D.; Zhou, C.; Wang, Z. Long Short-Term Memory Neural Network with Transfer Learning and Ensemble Learning for Remaining Useful Life Prediction. Sensors 2022, 22, 5744. https://0-doi-org.brum.beds.ac.uk/10.3390/s22155744

AMA Style

Wang L, Liu H, Pan Z, Fan D, Zhou C, Wang Z. Long Short-Term Memory Neural Network with Transfer Learning and Ensemble Learning for Remaining Useful Life Prediction. Sensors. 2022; 22(15):5744. https://0-doi-org.brum.beds.ac.uk/10.3390/s22155744

Chicago/Turabian Style

Wang, Lixiong, Hanjie Liu, Zhen Pan, Dian Fan, Ciming Zhou, and Zhigang Wang. 2022. "Long Short-Term Memory Neural Network with Transfer Learning and Ensemble Learning for Remaining Useful Life Prediction" Sensors 22, no. 15: 5744. https://0-doi-org.brum.beds.ac.uk/10.3390/s22155744

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop