Next Article in Journal
MiNgMatch—A Fast N-gram Model for Word Segmentation of the Ainu Language
Previous Article in Journal
A Comparison of Reinforcement Learning Algorithms in Fairness-Oriented OFDMA Schedulers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Temperature Forecast of Ship Propulsion Devices from Sensor Data

1
School of Maritime Economics and Management, Dalian Maritime University, Dalian 116026, China
2
Logistics Research Institute, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Submission received: 5 September 2019 / Revised: 9 October 2019 / Accepted: 14 October 2019 / Published: 16 October 2019

Abstract

:
The big data from various sensors installed on-board for monitoring the status of ship devices is very critical for improving the efficiency and safety of ship operations and reducing the cost of operation and maintenance. However, how to utilize these data is a key issue. The temperature change of the ship propulsion devices can often reflect whether the devices are faulty or not. Therefore, this paper aims to forecast the temperature of the ship propulsion devices by data-driven methods, where potential faults can be further identified automatically. The proposed forecasting process is composed of preprocessing, feature selection, and prediction, including an autoregressive distributed lag time series model (ARDL), stepwise regression (SR) model, neural network (NN) model, and deep neural network (DNN) model. Finally, the proposed forecasting process is applied on a naval ship, and the results show that the ARDL model has higher accuracy than the three other models.

1. Introduction

In recent years, with the increase of global trade volume, seaborne transport volume shows a momentum of accelerating rise, and maritime transport has undoubtedly become a very important transport mode. However, the problems brought about by the maritime industry in terms of marine environment and ecological protection, energy sustainability, etc., have become increasingly serious. It is imperative for ships to save energy, reduce emissions, and minimize marine pollution by means of using new and cleaner energy as ship power, building ships with new materials, economically efficient hull design, economic speed, and intelligent ship fault prediction.
The ship propulsion system [1] is the heart of ship, and it is the basic guarantee for the safety of the ship operation. Once the fault of propulsion devices occurs, it will not only affect the normal operation of the entire ship’s power plants, but also cause a series of environmental problems, such as diesel oil leakage and seawater pollution. With the continuous improvement of scale and automation of modern ship propulsion systems, the design of various control systems is more and more complex, which makes it difficult to locate and solve the faults in time. However, many propulsion devices of ships are equipped with sensors, and the temperature change of the ship’s propulsion devices can be forecasted for identifying potential faults, which will then save the cost of operation and maintenance, reduce energy consumption, and then ensure the sustainable development of ship energy.
There are many references focusing on the analysis of ship sensing data. For example, Onwuegbuchunam et al. [2] conducted research and analysis on the marine pollution of ships in Nigeria, and the integrated model was proposed as an alternative administrative tool for monitoring and controlling pollution in seaports. Iodice et al. [3] proposed a numerical approach to assess air pollution by ship engines in maneuver mode and fuel switch conditions, and it was also used to assess the impact of marine engine pollutant emissions on ambient air quality in coastal areas. Kim and Park [4] proposed a method to consider the optimization of the hull form of ULCS (Ultra Large Container Ship) in green ship design, and confirmed that the energy efficiency of the optimized hull form was improved by the proposed total energy formula. Gudehus and Kotzab [5] developed a model for the economic travel speed of container ships, which showed how the travel speed impacts the profit situation as well as the environmental sustainability. The above references focus on data-driven methods to solve ship problems, but they are not fault or prediction problems. Meanwhile, there are some references for predicting the possible fault locations and causes, and providing maintenance decisions and suggestions, so as to take effective measures to quickly troubleshoot. For instance, Feng and Li [6] aimed at the inaccurate and time-consuming problems of the fault diagnosis method for a large-scale ship engine, and proposed an intelligent diagnosis method for large-scale ship engine fault in a non-deterministic environment based on neural network. Tang et al. [7] proposed a fault diagnosis system of ship power plants based on a C/S (Client/Server) and B/S (Browser/Server) hybrid structure for the ship power plant fault diagnosis system, which provided a good solution for the development of intelligent ships. Yang and Tang [8] proposed a method for fault pattern recognition and state prediction of marine power plants based on HMM-SVR. The above research provides a lot of prior knowledge for ship condition prediction and fault diagnosis. However, the data collected by various sensors is time-series data, and it is necessary to maintain its time-series characteristics. So, we need to consider its time characteristics in the research process as follows: Wu et al. [9] proposed a hybrid model to represent the tracking dynamic behavior of ships in order to achieve ship-tracking control along the desired path at constant speed. McCullough et al. [10] proposed constructing ordinal networks from discrete sampled continuous chaotic time series, and then regenerated new time series by taking random walks on the ordinal networks. Cinar et al. [11] combined the time series with the recurrent neural network (RNN), and proposed the period-aware content attention RNNs for time series forecasting with missing values. Yang and Liu [12] proposed a time-series prediction model based on the hybrid combination of high-order FCMs (HFCMs) with the redundant wavelet transform to handle large-scale non-stationary time series.
Therefore, we proposed a data-driven method for forecasting the temperature of ship propulsion devices, and the difference between true temperature and forecasting temperature can be used to evaluate the accuracy of prediction. At the same time, we compare four different methods, and the results show that the data have obvious linear characteristics. The proposed method can be divided into three aspects: (1) data preprocessing for removing records with errors and missing values; (2) feature selection for removing unrelated features; and (3) forecasting the temperatures. Finally, the model is applied on the sensing data from the propulsion devices of a naval ship, and four forecasting methods are compared to test which is the best.
The reminder of the paper is arranged as follows. In Section 2, the dataset is presented, and the preprocessing, feature selection, and four prediction models are described. The results are shown and discussed in Section 3. Finally, we conclude the paper in Section 4.

2. Materials and Methods

2.1. Datasets

The dataset adopted in this study is from the propulsion system of a naval ship, generated by a numerical simulator of a naval vessel [13]. This dataset (www.cbm.smartlab.ws) was released on the widespread well-known dataset repository of the University of California in Irvine (UCI) [14], which consists of 589,224 records with 30 features, including 25 features and 5 coefficients for ship-related features.

2.2. Data Preprocessing and Feature Selection

Since we want to analyze sensor data directly, 5 coefficient features were removed, and 25 features remained. These 25 features contain two temperature-related features: HPTET and GTCOAT, which are considered as the key factors for judging whether the fault occurs or not. Among the remaining 23 features, 6 features are single-valued features that have no effect on the results, and can be removed directly. The remaining 17 features are respectively referred to as Level, Speed, GTST, GTS, CPPTS, STP, SRP, GGS, FF, ABBTCS, GTCOAP, EP, HPTEP, TCSTCS, TCS, PRS, and PTP, which can be used as independent variables for predicting HPTET and GTCOAT. Subsequently, we normalize the values of these 17 features to fall in the range of 0 to 1, and predict objectives based on different samples. The HPTET is predicted by the H dataset, which consists of the HPTET temperature and 17 features, while GTCOAT is predicted by the G dataset, which consists of the GTCOAT temperature and 17 features. These data have time-series characteristics and we don’t know which model is more effective for forecasting the temperature, so we compare and employ several prediction models in this paper. The majority (70%) of the dataset was chosen as training data, and the remaining were test data.
Finally, seven feature selection methods, including linear regression (Lr) [15], ridge regression (Ridge) [16], lasso regression (Lasso) [17], random forest (RF) [18,19], correlation (Corr) [20], stability (Stability) [21], and the average of these six results (AVG) are employed for feature selection in the H training data and G training data (see Section 3).

2.3. Prediction Models

It is very difficult to select the most effective method from many prediction models, so we adopt the four most commonly used prediction models for getting the best one. In this study, four prediction models are employed: the autoregressive distributed lag time series model (ARDL) model, stepwise regression (SR) model, neural network (NN) model [22], and deep neural network (DNN) model, respectively.

2.3.1. ARDL Model

The observations on a measurable variable acquired over a period of time form a time series [23]. ARDL model is a time-series model that is commonly used for statistical and econometrical analysis, and can be used to describe the behavior of a dependent variable; Y t is the output, μ is a disturbance, ε t is an independent variable, X t j and X t i are lags before the i and j time units. The number of lags used in the ARDL model determines how many past units are considered to affect the dependent variable.
An ARDL time-series model can be expressed as shown in Formula (1):
Y t = μ + i = 1 p γ i X t i + j = 0 r β j X t j + ε t
where:
ε t ~ i , i , d   N ( 0 , σ ε 2 )
Here, p and r are the lag length for the dependent variable Y t and independent variable X t , respectively. The ARDL model can also be expressed as follows:
C ( L ) Y t = μ + B ( L ) X t + ε t
where:
C ( L ) = 1 γ 1 L γ 2 L 2 γ P L P
B ( L ) = β 0 + β 1 L + β 2 L 2 + + β r L r
Here, L refers to the lag operator [23]. In this study, we set L = 1, 5, 10, 15 for verifying and comparing the effects of lag on the results, which are 1 lag step, 5 lag steps, 10 lag steps, and 15 lag steps, respectively.
After regression prediction is employed for the corresponding coefficient on training data, the coefficient is verified by T-test with the tool of R language. Then, the verified coefficients are brought into the test set to obtain the predicted temperature.

2.3.2. Stepwise Regression Model

The stepwise regression algorithm is a widely used regression algorithm [24]. Its basic idea is to introduce the independent variables one by one, and the sequence of independent variables influences the dependent variable Y significantly. Each time, the old independent variables are verified one by one while a new independent variable is input. The insignificantly independent variables in the current equation are removed one by one from the independent variables that have the least impact on the dependent variable Y, until you can’t introduce new arguments. Finally, the independent variables retained in the regression equation all have a significant impact on the dependent variable Y, while independent variables that are not in the regression equation have an insignificant effect on Y. Such a regression equation is the optimal equation. The basic steps of stepwise regression can be described in Figure 1.
Assuming that there are m independent variables and n groups of observations, the linear regression model of m elements can be shown as Formula (6):
y i = β 0 + β 1 x i 1 + β j x i j + β m x i m + e i
where i = 1, 2, …, n represents the i-th observation, j = 1, 2, …, m represents the j-th independent variable, βj is the partial regression coefficient, and ei is the random error.
Firstly, the data is normalized, and the average values x j ¯ = 1 n i = 1 n x i j ,   y ¯ = 1 n i = 1 n y i , and the standard value σxj and σy are calculated; then, the correlation coefficient of xj and y are calculated as R = ( r i j ) M × M . Here, M is the number of all variables including independent variables, and the last row and last column are the correlation coefficients of respective variables and dependent variables. Each independent variable has a significant effect on the dependent variable, and the sum of the squared error of the partial regression is the largest. The inverse transform is used to transform the correlation coefficient matrix, and the previously selected independent variables are tested. If it is not significant, the independent variable will be removed. The process is repeated until all the introduced independent variables have a significant effect on the dependent variable. If k − 1 is reached and m − 1 independent variables are selected at the same time, the following four steps can be taken in step k.
First of all, the partial determination coefficient not selected from the independent variables is calculated according to Formula (7):
U p i ( k ) = [ r i m ( k 1 ) ] 2 r i i ( k 1 )
where i is an independent variable that has not yet been selected, and r i i represents the i column element of row i on the diagonal line of the correlation coefficient matrix.
Secondly, we compare and find the independent variable x1 with the largest partial determination coefficient using Formula (8):
U p 1 ( k ) = m a x ( U p i ( k ) )
Thirdly, we test the significance of partial regression x1, as shown in Formula (9):
F = U p 1 ( k ) ( r M M ( k 1 ) U P 1 ( K ) ) / ( M m 1 )
If the result of the F-test is not significant, then there is no x1 available for introduction, and the regression is terminated. On the contrary, if the result of the F-test is significant, x1 is selected, and the r(k−1) array is inversely transformed into r(k). The process is shown by Equations (10)–(13).
r u u ( k ) = 1 r u u ( k 1 )           ( i = u , j = u )
r u j ( k ) = r u j ( k 1 ) r uu ( k 1 )           ( i = u ,   j u )
r i u ( k ) = r i u ( k 1 ) r u u ( k 1 )           ( i u , j = u )
r i j ( k ) = r i j ( k 1 ) r i u ( k 1 ) r u j ( k 1 ) r u u ( k 1 )             ( i u , j u )
Last but not least, if x1 is selected, the partial determination coefficient of the previously selected independent variable x1 is the same as the value obtained in second step. The F-test is performed on the previously selected independent variable x1, as shown in Formula (14):
F = U p i ( k ) r M M ( R ) / ( M m 1 )
where r M M ( k ) is the sum of the square error of the standard deviation regression. For the selected x1, if the dependent variable is significant, we should leave the significant independent variable for the dependent variable, and the process jumps to the first step. If there is an insignificant existence, the independent variable with the smallest deviation coefficient is defined as x1, the r(k) matrix is re-transformed, and the program jumps to the first step. The steps mentioned above are repeated until all the significantly independent variables are introduced, and all the non-significantly independent variables are eliminated, so as to obtain the regression equation of the relationship between the independent variable and the dependent variable.

2.3.3. Neural Network Model

A neural network with a hidden layer is used in this study. The full horizontal connection is achieved throughout the network, while the neurons in each layer between the verticals are disconnected. The network is a feed-forward network trained by the error back-propagation algorithm under the minimum error between the actual output and the expected output of the network. The multi-layer connection weights and thresholds of the network are corrected from the back to the front layer. This error back-propagation correction continues, and the correct rate of response of the neural network to the input mode is also rising. The structure of the network is given in Figure 2.
When constructing the neural network model, a random value in the range of (−1,1) to each connection weight and threshold is first assigned; then, the vector P k = ( p 1 , p 2 , , p n ) and the expected output are input as the training samples. The vector Q k = ( q 1 , q 2 , , q i ) is provided to the network. Then, the input vector Pk, the connection weight v i j , and the threshold θ j are used to calculate the input of each unit of the hidden layer. Then, the input is used to calculate the output of each unit of the hidden layer by the transfer function f:
O j = f ( i = 1 n v i j p i θ j )
where i is the dimension of the input layer i = (1, 2, ..., n). Then, the output O j is used to connect weights ω j t with ψ t thresholds of the hidden layer to calculate the output of each unit of the hidden layer. Then, we use the output of the hidden layer to calculate the output of the output layer through the transfer function f; the actual output of the network is given in Formula (16):
Y t = f ( j = 1 p ω j t O j ψ t )
where j is the j-th dimension of the hidden layer. In Equations (15) and (16), the transfer function f is usually expressed by a sigmoid-type function:
f ( x ) = 1 1 + e x
The expected output vector O i and the actual output vector Yt are used to calculate the error of the output layer.
d t k = ( q t k Y t ) f ( j = 1 p ω j t O j ψ t )
where f’ is the derivative of the output layer function. Then, the connection weight ω j t and the output of each unit of the hidden layer are used to calculate the correction error of the hidden layer.
e j k = ( t = 1 q ω j t d t ) f ( i = 1 n v i j p i θ j )
After obtaining the above correction error, the error is used to correct the connection weight and threshold between the output layer and the hidden layer and between the hidden layer and the input layer. The correction amounts are:
Δ ω j t = α d t k O j
Δ ψ t = α d t k
where 0 < α < 1.
Δ v i j = β e j k p i k
Δ θ j = β e j k
where 0 < β < 1.
We used the TensorFlow tool to implement the models in this paper; a three-layer multi-input and single-output neural network with one hidden layer is adopted to establish a prediction model. However, if the number of neurons in the hidden layer is very large, the calculating amount will increase, and the over-fitting problem will easily occur. If the number of neurons is too small, the network performance will be affected, and the expected effect will not be achieved. Therefore, it is necessary to have a reasonable number of neurons. Through many experiments, we finally selected 20 neurons in the hidden layer, and employed the momentum method to define the optimizer and get the prediction results.
Meanwhile, we constructed a deep neural network (DNN) model because it often provides better performance than general neural networks with one hidden layer. The DNN model adopted in this paper is implemented by TensorFlow, and is composed of five hidden layers and one output layer.

3. Results and Discussion

3.1. Preprocessing and Feature Selection

In this paper, features are selected by their contribution, which is calculated by seven feature selection methods on the H dataset and G dataset mentioned above, which are chosen in descending order and in accordance with the rule that the sum of contribution values of selected features is over 95% (see Table 1 and Table 2). The other features whose total sum of contribution values is less than 5% are considered to have a lesser influence on the results, and can be ignored.
Here, a1–a17 represent 17 features of Level, Speed, GTST, GTS, CPPTS, STP, SRP, GGS, FF, ABBTCS, GTCOAP, EP, HPTEP, TCSTCS, TCS, PRS, and PTP, respectively. The data marked in bold in the tables is the final selected variable attribute corresponding to it.
The process of feature selection and data processing for the HPTET temperature are same to that of the GTCOAT temperature, which is given in Table 2.

3.2. Results of GTCOAT Temperature Prediction.

In the present section, we predicted the test dataset according to the correlation coefficient obtained from the training data experiment, and compared the predicted results with the actual data. Three error analysis methods were employed, which are the actual error, absolute error, and relative error. Then, we drew two kinds of error graph, namely the actual error and relative error. Meanwhile, we compared their mean absolute error (MAE) and mean average relative error (MARE), and comprehensively analyzed the prediction effect. By comparing the seven feature selection methods for the time series, we found that for the GTCOAT temperature, the AVG feature selection method had the best prediction results (see Table 3). Moreover, the ARLD model with 15 lag steps has the smallest prediction error among all the models, whose MAE value is 1.0491 and MARE value is 0.0016. Table 3 shows that the DNN model is more accurate than the NN model, so the prediction accuracy can be improved by increasing the number of hidden layers. However, at the same time, it brings many problems, such as over-fitting and so on.
The data in Table 3 marked in black bold color is the data obtained under the best selected features of the final selection. The abbreviations for the following prediction methods make the table simpler and clearer.
  • LR1: Time-Series Model with 1 lag.
  • NN: Neural Network Model.
  • DNN: Deep Neural Network Model.
  • SR: Stepwise Regression Model.
Due to the large amount of data in this experiment, we have to select continuous data to visualize the results. Figure 3 and Figure 4 respectively show the actual error and relative error between the predicted and actual values of all models. Figure 3 shows that the actual error range of the ARDL model with 15 lag steps is the smallest, most of which are concentrated between –3 and 3. At the same time, Figure 4 shows that the relative error of this model is the smallest, mostly less than 0.06. Therefore, we can conclude that the ARDL model with 15 lag steps has the best results for predicting the GTCOAT temperature.

3.3. Results of HPTET Temperature Prediction.

Similarly, the process of models for predicting the HPTET temperature equals to that for predicting the GTCOAT temperature. The results of all models for predicting the HPTET temperature are shown in Table 4, which shows that the MAE and the MARE of the ARDL model with 15 lag steps is the smallest after the AVG feature selection method. The MAE and MARE of this model are 2.5255 and 0.0037, respectively.
Subsequently, we also select 1000 continuous records to visualize the results. Figure 5 and Figure 6 showed the actual error and relative error of all the models for predicting the HPTET temperature. Figure 5 shows that the actual error range of the ARDL model with 15 lag steps is the smallest, most of which are concentrated between –10 and 10. At the same time, from Figure 6, we can easily find that the relative error of this model is the smallest, mostly less than 0.02. Therefore, we can conclude that the ARDL model with 15 lag steps has the best prediction results for predicting the HPTET temperature.

3.4. Discussions

Based on the above experimental results, we know that the ARDL model with the AVG feature selection method shows better results for forecasting both the GTCOAT and HPTET temperatures. For the GTCOA temperature, when the ARDL model has one lag step, the AVG feature selection method shows that the MAE is 2.3686 and an MARE of 0.0037. With 5 lag steps, the MAE is 1.3055 and the MARE is 0.0020. With 10 lags, the MAE is 1.0542 and the MARE is 0.0016. With 15 lags, the MAE is 1.0491 and the MARE is 0.0016. Among the rest of the models, the MAE and MARE of the multiple linear regression method are 2.4770 and 0.0039, respectively. The MAE and MARE of the stepwise regression model are 2.7024 and 0.0042, respectively. The MAE and MARE of the NN model are 9.3936 and 0.0141, respectively. The MAE and MARE of the DNN model with the Lasso feature selection method are 33.5700 and 0.0502, respectively.
For the HPTET temperature, when the ARDL model has one lag step, the MAE and the MARE are 3.3385 and 0.0049, respectively. With 5 lag steps, the MAE and the MARE are 2.8086 and 0.0041, respectively. With 10 lags, the MAE and MARE are 2.5311 and 0.0037, respectively. With 15 lags, the MAE and MARE are 2.5255 and 0.0037, respectively. Among the rest of the models, the MAE and MARE of the linear regression model are 3.3548 and 0.0049, respectively. The MAE and MARE of the stepwise regression model are 20.8675 and 0.0317, respectively. The MAE and MARE of the NN model are 66.4863 and 0.08556, respectively. The MAE and MARE of the DNN model with the Corr feature selection method are 15.5496 and 0.0233, respectively.
According to the analysis mentioned above, we know that the ARGL model with 15 lags under AVG feature selection has the smallest error and the best prediction accuracy. The reason is that the remote sensing data from ships is time series, which means that the time-series correlation prediction model is more effective. In addition, we analyzed the running time of these models. Our experiments are carried out on a 64-bit Windows 7 operation system computer, in which the CPU is an Intel(R) Core(TM) i7-3370 cpu @ 3.40GHz 3.40GHz. All the models are sorted by the running time in ascending order: The LR model takes about 55 seconds, the ARDL models with 1 lag step, 5 lag steps, 10 lag steps, and 15 lag steps respectively take about 60 seconds, 75 seconds, 95 seconds, and 110 seconds; the NN model takes about 460 seconds; lastly, the DNN model takes about 1500 seconds, which shows that the performance of the neural network model is worse than the ARDL model, and the more complex the network is, the more time it takes.
Some potential failures of the propulsion system can be found by the difference between the actual temperature and the predicted temperature, since sudden temperature changes are mainly caused by device failure. In actual voyage, it not only helps the crew to identify the location of the faults, but also automatically reminds them to focus on warnings of possible failures. The maintenance personnel only need to locate the specific parts of the abnormal situation in time and carry out inspection and maintenance, which can greatly reduce the workload of the crew and reduce the operation and maintenance cost of the ship.

4. Conclusions

Timely potential fault prognostics and early warnings in advance can effectively reduce the cost of operation and maintenance of ships, thereby further saving energy and reducing the cost of the operation and maintenance of ships. Device failure is often accompanied by sudden changes in temperature. Therefore, we focused on the temperature forecast of ship propulsion devices in this paper, and based on the states and position information of the ship transmitted to the shore at regular intervals, carried out a series of analysis and mining on the collected data, and constructed the several forecasting models. It is essential to compare these models because we lack prior knowledge, and the results can be summarized as follows:
(1) Compared with other feature selection methods, the AVG feature selection method shows the best effect.
(2) When forecasting the GTCOAT temperature, the MAE values of the ARDL model with 15 lag steps, the linear regression model, the stepwise regression model, the NN model, and the DNN model are 1.0491, 2.4770, 2.7024, 9.3936, and 15.5496, respectively while the MARE values of these models are 0.0016, 0.0039, 0.0042, 0.0141, and 0.0233 respectively. Similarly, when forecasting the HPTET temperature, the MAE values of the ARDL model with 15 lags, the linear regression model, the stepwise regression model, the NN model, and the DNN model are 2.5255, 2.4770, 20.8675, 66.4863, and 33.5700, respectively, while the MARE values of these models are 0.0037, 0.0039, 0.0317, 0.0855, and 0.0502, respectively.
The results show that the temperature forecast of the ship propulsion devices displays the linear features in two aspects: One is that the ARDL model has the most stable results and the highest accuracy. Another is that the NN model often shows better results for solving non-linear problems, but in this case, it shows worse results than the two other models.
(3) The forecasted temperatures can remind crew on board to discover potential failures timely, which will contribute to correct maintenance decisions and suggestions, so as to take effective measures to quickly troubleshoot the faults. Consequently, it is conducive to energy conservation, emission reduction, environmental protection, and sustainable energy development.
In this paper, we only predict the temperature of the ship propulsion system, but do not analyze the acceptable margin of error and how to identify the fault of ship propulsion devices because of the lack of empirical knowledge. In the future, we will try our best to improve the precision, reduce the training time, and discuss the reasonable margin of error.

Author Contributions

Conceptualization: T.L. and M.H.; methodology: T.L.; validation: Q.Y. and M.H.; formal analysis: Q.Y.; investigation: T.L.; resources: M.H.; data curation: T.L.; writing—original draft preparation: M.H.; writing—review and editing: T.L.; visualization: M.H.; supervision: T.L.; project administration: T.L.; funding acquisition: T.L.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 71271034, the National Social Science Foundation of China, grant number 15CGL031, the Fundamental Research Funds for the Central Universities, grant numbers 3132019353 and 3132019028, the Program for Dalian High Level Talent Innovation Support, grant number 2015R063, and the National Natural Science Foundation of Liaoning Province, grant number 20180550307.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ji, Y.L.; Sun, Y.Q.; Zhang, H.P.; Zhang, Y.D.; Chen, H.Q. Novel Ship Propulsion System. Chin. J. Mech. Engin. 2009, 22, 198–206. [Google Scholar] [CrossRef]
  2. Onwuegbuchunam, D.E.; Ebe, T.E.; Okoroji, L.I. An Analysis of Ship-Source Marine Pollution in Nigeria Seaports. J. Mar. Sci. Eng. 2017, 5, 39. [Google Scholar] [CrossRef]
  3. Iodice, P.; Langella, G.; Amoresano, A. A numerical approach to assess air pollution by ship engines in manoeuvring mode and fuel switch conditions. Energy Environ. 2017, 28, 827–845. [Google Scholar] [CrossRef]
  4. Kim, M.; Park, D.W. A study on the green ship design for ultra large container ship. J. Mar. Sci. Eng. 2015, 21, 558–570. [Google Scholar] [CrossRef]
  5. Gudehus, T.; Kotzab, H. Economic Ship Travel Speed and Consequences for Operating Strategies of Container Shipping Companies. In Proceedings of the 4th International Conference on Dynamics in Logistics (LDIC), Bremen, Germany, 10–14 February 2014; pp. 407–415. [Google Scholar]
  6. Feng, D.H.; Li, Y.H. Research on Intelligent Diagnosis Method for Large-Scale Ship Engine Fault in Non-Deterministic Environment. Pol. Marit. Res. 2017, 24, 200–206. [Google Scholar] [CrossRef] [Green Version]
  7. Tang, G.G.; Fu, X.D.; Chen, N. Ship Power Plant Remote Fault Diagnosis System Based on B/S and C/S Architecture. Ship Eng. 2018, 40, 66–71. [Google Scholar]
  8. Yang, Y.F.; Feng, J. Fault Pattern Recognition and State Prediction Research of Ship Power Equipment Based on HMM-SVR. Ship Eng. 2018, 40, 68–72, 97. [Google Scholar]
  9. Wu, J.P.; Peng, H.; Ohtsu, K. Ship’s tracking control based on nonlinear time series model. Appl. Ocean Res. 2012, 36, 1–11. [Google Scholar] [CrossRef]
  10. McCullough, M.; Sakellariou, K.; Stemler, T.; Stemler, T. Regenerating time series from ordinal networks. Chaos. 2017, 27. [Google Scholar] [CrossRef]
  11. Cinar, Y.G.; Mirisaee, H.; Goswami, P. Period-aware content attention RNNs for time series forecasting with missing values. Neurocomput. 2018, 312, 177–186. [Google Scholar] [CrossRef]
  12. Yang, S.; Liu, J. Time-Series Forecasting Based on High-Order Fuzzy Cognitive Maps and Wavelet Transform. IEEE Trans. Fuzzy Syst. 2018, 26, 3391–3402. [Google Scholar] [CrossRef]
  13. Coraddu, A.; Oneto, L.; Ghio, A. Machine learning approaches for improving condition- based maintenance of naval propulsion plants. Proc. Inst. Mech. Eng. Part M J. Eng. Marit. Environ 2016, 230, 136–153. [Google Scholar] [CrossRef]
  14. Cipollini, F.; Oneto, L.; Coraddu, A.; Murphy, A.J.; Anguita, D. Condition-based maintenance of naval propulsion systems with supervised data analysis. Ocean Eng. 2018, 149, 268–278. [Google Scholar] [CrossRef]
  15. Yerel, S.; Ersen, T. Prediction of the Calorific Value of Coal Deposit Using Linear Regression Analysis. Energy Sour. Part A Recov. Util. Environ. Eff. 2013, 35, 976–980. [Google Scholar] [CrossRef]
  16. Liu, X.Q.; Gao, F.; Xu, J.W. Linearized Restricted Ridge Regression Estimator in Linear Regression. Commun. Stat.-Theory Methods 2012, 41, 4503–4514. [Google Scholar] [CrossRef]
  17. Chang, L.; Roberts, S.; Welsh, A. Robust Lasso Regression Using Tukey’s Biweight Criterion. Technometrics 2018, 60, 36–47. [Google Scholar] [CrossRef]
  18. Scornet, E. Random Forests and Kernel Methods. IEEE Trans. Inf. Theory 2016, 62, 1485–1500. [Google Scholar] [CrossRef] [Green Version]
  19. Díaz-Uriarte, R.; DeAndres, S.A. Gene selection and classification of microarray data using random forest. BMC Bioinform. 2006, 7, 1–13. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Liu, T.; Li, K.F.; Zhang, J.W. Improved visual correlation analysis for multidimensional data. J. Vis. Lang. Computing 2017, 41, 121–132. [Google Scholar]
  21. Wen, M.L.; Qin, Z.F.; Kang, R. Sensitivity and stability analysis in fuzzy data envelopment analysis. Fuzzy Optim. Decis. Mak. 2011, 10, 1. [Google Scholar] [CrossRef]
  22. Acikkar, M.; Sivrikaya, O. Prediction of gross calorific value of coal based on proximate analysis using multiple linear regression and artificial neural networks. Turk. J. Electr. Eng. Comput. Sci. 2018, 26, 2541–2552. [Google Scholar] [CrossRef]
  23. Yue, S.P. Modeling of Engine Parameters for Condition-Based Maintenance of the MTU Series 2000 Diesel Engine. Master’s Thesis, Naval Postgraduate School, Monterey, CA, USA, September 2016. [Google Scholar]
  24. Agostinelli, C. Robust stepwise regression. J. Appl. Stat. 2002, 29, 825–840. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Steps of the stepwise regression algorithm.
Figure 1. Steps of the stepwise regression algorithm.
Information 10 00316 g001
Figure 2. The structure of the neural network.
Figure 2. The structure of the neural network.
Information 10 00316 g002
Figure 3. Actual error of all models for predicting the GTCOAT temperature. Figure 3 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Figure 3. Actual error of all models for predicting the GTCOAT temperature. Figure 3 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Information 10 00316 g003aInformation 10 00316 g003b
Figure 4. Relative error of all models for predicting the GTCOAT temperature. Figure 4 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Figure 4. Relative error of all models for predicting the GTCOAT temperature. Figure 4 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Information 10 00316 g004aInformation 10 00316 g004b
Figure 5. Actual error of ARDL model for predicting the HPTET temperature. Figure 5 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Figure 5. Actual error of ARDL model for predicting the HPTET temperature. Figure 5 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Information 10 00316 g005aInformation 10 00316 g005b
Figure 6. Relative error of ARDL model for predicting the HPTET temperature with different lag steps. Figure 6 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Figure 6. Relative error of ARDL model for predicting the HPTET temperature with different lag steps. Figure 6 contains eight graphs, which were listed as follows: (a) ARDL model with lag 1, (b) ARDL model with lag 5, (c) ARDL model with lag 10, (d) ARDL model with lag 15, (e) LR model, (f) NN model, (g) SR model, (h) DNN model.
Information 10 00316 g006aInformation 10 00316 g006b
Table 1. Contribution of features by different selection methods for the G dataset.
Table 1. Contribution of features by different selection methods for the G dataset.
FeaturesLinear
Regression
Ridge
Regression
Lasso
Regression
Random ForestCorrelationStabilityAVG
a10.02400.0071 0.0248 0.0022 0.05290.25180.0605
a20.01020.06080.09430.0019 0.04890.10270.0531
a30.00610.10910.0000 0.0007 0.07460.04890.0399
a40.09280.04020.10690.0150 0.06970.04400.0614
a50.00020.05200.0000 0.01680.05540.0073 0.0219
a60.31790.0031 0.0000 0.0000 0.06880.0122 0.0670
a70.01210.04020.0000 0.10650.06970.02450.0422
a80.00230.03510.26300.0080 0.0256 0.06850.0671
a90.00840.10840.29950.57730.03140.0000 0.1708
a100.00970.10170.0000 0.09160.03130.0000 0.0391
a110.00030.11200.21150.08280.15710.26890.1388
a120.03710.19860.0000 0.0025 0.05670.0049 0.0500
a130.05140.03980.0000 0.0004 0.10190.12220.0526
a140.00120.0046 0.0000 0.08270.0199 0.0024 0.0185
a150.00150.0007 0.0000 0.0000 0.0026 0.0000 0.0008
a160.09520.04000.0000 0.0116 0.06970.02200.0398
a170.32960.04670.0000 0.0000 0.06390.0196 0.0766
Table 2. Contribution of features by different selection methods for the H dataset.
Table 2. Contribution of features by different selection methods for the H dataset.
FeaturesLinear
Regression
Ridge
Regression
Lasso
Regression
Random ForestCorrelationStabilityAVG
a10.01650.01470.00000.0000 0.02440.07480.0217
a20.0057 0.01730.01590.0000 0.0230 0.02620.0147
a30.12620.19300.23020.0000 0.12180.20370.1458
a40.02410.0067 0.05500.0000 0.04860.0037 0.0230
a50.0025 0.03450.0000 0.0000 0.06800.0093 0.0191
a60.29520.12790.06170.0000 0.08310.0131 0.0968
a70.07630.0066 0.0027 0.0000 0.04860.0019 0.0227
a80.0074 0.01530.04630.0000 0.0151 0.13640.0368
a90.05010.17550.25630.92210.08730.20560.2828
a100.0041 0.0001 0.09590.0000 0.05930.02060.0300
a110.04140.09520.02270.0000 0.09960.20370.0771
a120.05650.11540.19820.00120.05190.0000 0.0705
a130.0092 0.06240.0000 0.0000 0.09830.04860.0364
a140.0008 0.0005 0.0000 0.07650.03900.04860.0276
a150.0003 0.0093 0.0053 0.0000 0.0024 0.0000 0.0029
a160.10100.0066 0.0000 0.0000 0.04860.0000 0.0261
a170.18260.11920.0099 0.0000 0.08090.0037 0.0660
Table 3. Results of models for forecasting the GTCOAT temperature. ARDL: autoregressive distributed lag time series model, DNN: deep neural network, LR1: time-series model with 1 lag, MAE: mean absolute error, MARE: mean average relative error, NN: neural network, SR: stepwise regression.
Table 3. Results of models for forecasting the GTCOAT temperature. ARDL: autoregressive distributed lag time series model, DNN: deep neural network, LR1: time-series model with 1 lag, MAE: mean absolute error, MARE: mean average relative error, NN: neural network, SR: stepwise regression.
FeatureError
Measures
ARDLLRNNDNNcSR
LR1LR5LR10LR15
AVGMAE
MARE
2.3686
0.0037
1.3055
0.0020
1.0542
0.0016
1.0491
0.0016
2.4770
0.0039
9.3936
0.0141
17.9556
0.0271
2.7024
0.0042
CorrMAE
MARE
3.3785
0.0050
3.1014
0.0046
2.7143
0.0040
2.7087
0.0039
3.4136
0.0051
52.1559
0.0683
17.9556
0.0271
LassoMAE
MARE
3.4321
0.0050
2.9558
0.0043
2.5779
0.0037
2.5703
0.0037
3.4511
0.0050
53.3080
0.0699
15.5496
0.0233
LrMAE
MARE
3.6609
0.0053
3.3412
0.0049
2.7627
0.0040
2.7579
0.0040
3.7152
0.0054
46.1879
0.0575
20.2862
0.0306
RFMAE
MARE
20.9895
0.0312
12.5970
0.0183
6.0679
0.0088
6.0698
0.0088
21.8452
0.0328
41.7166
0.0572
28.8141
0.0460
RidgeMAE
MARE
3.3291
0.0049
2.9402
0.0043
2.5979
0.0038
2.5888
0.0038
3.3844
0.0049
21.3081
0.0335
27.7896
0.0405
StabilityMAE
MARE
7.7299
0.0111
5.4886
0.0077
4.9152
0.0069
4.9091
0.0068
8.5276
0.0122
47.2233
0.0600
17.8013
0.0268
Table 4. Results of models for forecasting the HPTET temperature.
Table 4. Results of models for forecasting the HPTET temperature.
FeatureError
Measures
ARDLLRNN
DNN
SR
LR1LR5LR10LR15
AVGMAE
MARE
3.3385
0.0049
2.8086
0.0041
2.5311
0.0037
2.5255
0.0037
3.3548
0.0049
66.4863
0.0855
50.5350
0.0645
20.8675
0.0317
CorrMAE
MARE
3.3785
0.0050
3.1014
0.0046
2.7143
0.0040
2.7087
0.0039
3.4136
0.0051
52.1559
0.0683
33.5700
0.0502
LassoMAE
MARE
3.4321
0.0050
2.9560
0.0043
2.5779
0.0037
2.5703
0.0037
3.4511
0.0050
53.3080
0.0699
54.1260
0.0700
LrMAE
MARE
3.6609
0.0053
3.3412
0.0049
2.7627
0.0040
2.7579
0.0040
3.7152
0.0054
46.1879
0.0575
38.9521
0.0615
RFMAE
MARE
20.9895
0.0312
12.5970
0.0183
6.0679
0.0089
6.0699
0.0088
21.8452
0.0328
41.7166
0.0572
326.0747
0.4034
RidgeMAE
MARE
3.3290
0.0049
2.9402
0.0043
2.5979
0.0038
2.5888
0.0038
3.3844
0.0049
21.3081
0.0335
42.6881
0.0575
StabilityMAE
MARE
7.7299
0.0111
5.4886
0.0077
4.9152
0.0069
4.9091
0.0069
8.5276
0.0122
47.2233
0.0600
48.1805
0.0612

Share and Cite

MDPI and ACS Style

Li, T.; Hua, M.; Yin, Q. The Temperature Forecast of Ship Propulsion Devices from Sensor Data. Information 2019, 10, 316. https://0-doi-org.brum.beds.ac.uk/10.3390/info10100316

AMA Style

Li T, Hua M, Yin Q. The Temperature Forecast of Ship Propulsion Devices from Sensor Data. Information. 2019; 10(10):316. https://0-doi-org.brum.beds.ac.uk/10.3390/info10100316

Chicago/Turabian Style

Li, Taoying, Miao Hua, and Qian Yin. 2019. "The Temperature Forecast of Ship Propulsion Devices from Sensor Data" Information 10, no. 10: 316. https://0-doi-org.brum.beds.ac.uk/10.3390/info10100316

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop