Next Article in Journal
Robust Smartphone Assisted Biosensing Based on Asymmetric Nanofluidic Grating Interferometry
Previous Article in Journal
Flexible Micro-Battery for Powering Smart Contact Lens
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Convolution Component-Based Method with Attention Mechanism for Travel-Time Prediction

1
School of Computer Science and Technology, University of Science and Technology Beijing, Beijing 100083, China
2
Informatization and Industry Development Department, State Information Center, Beijing 100045, China
3
College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China
4
Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
*
Authors to whom correspondence should be addressed.
Submission received: 8 April 2019 / Revised: 24 April 2019 / Accepted: 29 April 2019 / Published: 3 May 2019
(This article belongs to the Section Sensor Networks)

Abstract

:
Deep learning approaches have been recently applied to traffic prediction because of their ability to extract features of traffic data. While convolutional neural networks may improve the predictive accuracy by transiting traffic data to images and extracting features in the images, the convolutional results can be improved by using the global-level representation that is a direct way to extract features. The time intervals are not considered as aspects of convolutional neural networks for traffic prediction. The attention mechanism may adaptively select a sequence of regions and only process the selected regions to better extract features when aspects are considered. In this paper, we propose the attention mechanism over the convolutional result for traffic prediction. The proposed method is based on multiple links. The time interval is considered as the aspect of attention mechanism. Based on the dataset provided by Highways England, the experimental results show that the proposed method can achieve better accuracy than the baseline methods.

1. Introduction

In recent years, the increase in vehicle transit and congestion on highways and urban road networks has led the changes on traffic conditions to uncertainty. Traffic prediction is still a key functional component and a research focus has been on intelligent transportation systems (ITS). Accurate and timely traffic prediction can help individual travelers, business sectors and government agencies make traffic decisions. These decisions may alleviate traffic congestion, reduce NOx compounds emissions and improve traffic operation efficiency. In various traffic indexes including traffic volume, travel-time, average speed, queue length and the severity of incidents, the delivery of travel-time is widely accepted as an index of ITS because travel-time is very intuitive and is easily understood [1,2,3].
Deep learning approaches, such as the recurrent state–space neural network (SSNN) [4], the long short-term memory neural network (LSTM NN) [5], the autoencoders (AES) [6], the deep learning approach with a sequence of tanh layers [7] and CNN-based methods [8,9,10] have been proposed for traffic prediction in recent years with their abilities to extract effectively features from traffic data. Experimental results show that these deep learning approaches achieve superior performance over other baseline methods. The generalization ability of CNN is implemented by providing the constraints in a task domain including local receptive fields, shared weights and spatial or temporal subsampling [11]. CNN-based methods of traffic prediction may improve the predictive accuracy by transiting traffic data to images and extracting features of the images. The task of traffic prediction differs from image processing task by providing constrains to the convolution component. The constrains in image processing task are the relationship between adjacent pixels, the constrains in traffic prediction task are the relationship between each historical traffic data point and the predicted value [12]. The convolution of the existing CNN-based methods for traffic prediction is implement by convoluting the adjacent pixels of traffic data [8,9,10]. While the global-level representation was proposed to provide the constrains in the task domain and directly capture the relationship [12], the experiments are only based on the single link but not based on multiple links. Meanwhile, the time intervals are not considered as aspects of traffic prediction in convolutional neural networks.
The attention mechanism has recently demonstrated success in a wide range of tasks by concentrating on the differences between input features to better extracting features when different aspects are considered [13,14,15]. The attention mechanism is usually implemented by constructing a glimpse network [13] that is a two-layer neural network. The attentive neural network is trained during the training of the proposed model.
In this paper, we proposed the convolution component-based method with attention mechanism for travel-time prediction. The proposed method used the global level representation based on multiple links. For example, Figure 1 is a section of the road network. If the link AL282 will be predicted, the multiple links includes link AL282, AL283, AL2292, AL2274 and AL286. The time intervals are considered as the aspect of attention mechanism. Based on the dataset provided by Highways England, the proposed method is trained by using the back-propagation method and AdaGrad method to update the parameters of the proposed method. The experimental results show that the proposed method can achieve better predictive accuracy over baseline methods.
The main contributions of this paper can be summarized as follows.
  • To directly extract features in multiple links, we substitute the global-level representation for the traditional local receive fields in the input sequences.
  • We proposed the convolution component-based method with attention mechanism for travel-time prediction. The interval times are considered as the aspects of attention mechanism.
  • Based on the dataset provided by Highways England, the proposed method is trained and the experimental results show that the proposed method can achieve better predictive accuracy over the baseline methods.
The rest of this paper is organized as follows. Section 2 describes the related works for traffic prediction. Section 3 describes the proposed method for travel-time prediction. Section 4 described the dataset provided by Highways England and the experimental setting based on the dataset. Section 5 shows and discusses the experimental results. Section 6 summarizes the conclusions and future work.

2. Related Work

Traffic prediction is still a key functional component and a research focus of ITS. Various traffic prediction methods have been proposed and improved the prediction accuracy in recent years. Deep learning approaches have been proposed for traffic prediction because of effectively extracting the features of traffic data. CNN-based methods may improve the predictive accuracy by transiting traffic conditions to images and effectively extracting features in the images. Despite the success of the CNN-based methods, such methods are clearly limited because they do not directly extract the features in multiple links and do not use interval times as an aspect.

2.1. Traffic Prediction Methods

Various methods have been proposed for traffic prediction in the past years. The target functions of these methods relate the explicative variables to the target variable, which are usually implemented by using statistical and machine learning techniques [4]. According to the implementation techniques of these methods, these methods can be grouped into two categories including statistical methods and machine learning methods. Time series models are the most typical statistical methods and usually include the autoregressive (AR), moving average (MA), autoregressive moving average (ARMA), autoregressive integrated moving average (ARIMA), seasonal autoregressive integrated moving average (SAIMA) and Kalman filtering method. The machine learning methods for traffic prediction usually includes the linear regression (LR), k-nearest neighbor regression (k-NNR), support vector regression (SVR), SSNN, LSTM NN, AES and CNN.
ARIMA (p, d, q) model combines AR model and MA model, where parameters p, d, and q are non-negative integers, p is the order (number of time lags) of the autoregressive model, d is the degree of differencing (the number of times the data have had past values subtracted), and q is the order of the moving-average model. ARIMA model is a generalization of ARMA model. The I in ARIMA model is the difference of the observation values to make the time series stationary. In [16], the ARIMA(0, 1, 3) was proposed for short-term traffic prediction. Based on the 166 data sets from three surveillance systems that are deployed on a freeway in Los Angeles, the experimental results show that the proposed model achieves better performance than MA model. In [17], an ARIMA (0, 1, 2) was proposed for traffic prediction. Based on the dataset on five major urban arterials, the experimental results show that the proposed model is effective in reproducing the original time series.
SARIMA model is ARIMA model with a seasonal component. In [18], Wold decomposition theorem states that any stationary time series can be decomposed into a deterministic series and a stochastic series. Based on the Wold decomposition theorem, it is hypothesized that a weekly seasonal difference between traffic data can yield a weakly stationary transformation and univariate traffic data streams can be modeled as a SARIMA process. The experiments are performed based on the dataset from two freeways and validated the theoretical hypothesis.
In [19], a prediction scheme based on Kalman filtering technique was proposed for traffic flow prediction. The method uses both historic data (previous two days flow data). Real time data on the day of interest was also attempted. Promising results were obtained with mean absolute percentage error (MAPE) of 10 between observed and predicted flows.
Researchers in the traffic prediction field have paid much attention to machine learning methods in recent years because of their ability of extracting features in traffic data.
In [20], it was proposed for travel-time prediction that a linear relation (LR) between the predicted travel-time T T e ( t + δ ) and two naive predictors including the current status travel-time T e ( t ) and the historical mean travel-time μ T T ( t + δ ) . Based on the dataset from 116 single-loop detectors, the proposed LR method outperforms the principal components method and the k-NNR method.
The accuracy of k-NNR may be improved by using a larger dataset [21]. K-NNR methods have some advantages over SARIMA model [22] and can avoid current time series data to lead inefficient predictions [23].
Support vector regression machine does not depend on the dimension of the input vectors space and may nonlinearly map input vectors to a high-dimension feature space to construct a linear decision surface. Thus, SVR will have advantages in high dimensionality space [24]. In [25], an SVR method with a radial basis function kernel (RBF) was proposed for traffic prediction and achieves better performance than the current-time predictor and the historical-mean predictor based on a highway traffic dataset. In [26], the online version of SVR was proposed for short-term travel-time prediction and achieves better performance than the Gaussian maximum likelihood, Holt exponential smoothing and the artificial neural network. In [27], the incremental SVR method was proposed for traffic flow prediction and achieves better performance than the back-propagation neural network.
Various neural networks have been proposed for traffic prediction because of their ability to extract features in traffic data. The advantages and disadvantages of individual deep learning methods are shown in Table 1. In [28], the Elman recurrent neural networks [29] were referred to as SSNN for travel-time prediction based on the traffic state-space formulation. In [6], AES method was proposed for traffic flow prediction and was trained in a greedy layerwise fashion. Based on freeway data from the freeway system in California, the experimental results show that the AES method outperforms the random walk (RW), SVR, RBF network and back-propagation neural network for traffic flow prediction. In [5], LSTM NN was proposed for traffic speed prediction. Based on the dataset on an expressway without signal controls, LSTM NN outperforms the ARIMA(2, 2, 2), SVM, Kalman filter [3], Elman NN, time-delay neural network (TDNN) [30] and nonlinear autoregressive with exogenous inputs neural network (NARX) [31]. In [7], the deep learning approach combines a sequence of tanh layers and a linear layer to capture the impacts of breakdowns, recoveries or congestion on traffic flow prediction. Based on the dataset from twenty-one loop detectors, the experimental results show that the deep learning method is effective for traffic flow prediction.
In [8,9,10], CNN-based methods were proposed for traffic prediction by transiting traffic data to images and extracting the adjacent relationship in the images. In [8], the CNN-based method was proposed for traffic speed prediction. In [9], the fusion of CNN and LSTM was proposed for passenger demand prediction. In [10], the CNN-based method with an errorfeedback RNN was proposed for traffic speed prediction. In [12], based on the single link, the global-level representation was proposed to directly capture the relationship in single link.
While various traffic prediction methods have improved the accuracy of traffic prediction, these methods have been evaluated under different datasets and thus it is difficult to say that one method is clearly superior over the other methods under any traffic conditions [6].

2.2. Convolution Neural Network

To improve the learning ability of neural networks, a more interesting scheme is to rely on the topology of the input data [11]. CNN was proposed to implement the scheme by combining three architectural ideas including local receptive fields, shared weights and spatial or temporal subsampling. By applying feature mapping and weight sharing, the learning ability of neural networks of recognizing handwritten zip codes was enhanced [34]. In [11], LeNet-5 was proposed for document recognition and achieves better performance than the baseline methods.
Three architectural ideas can be seen as that CNN integrates the constraints in the task domain into its architecture. The idea of providing constraints in task domain to the proposed method has also been widely applied in many other fields. In [33], the convolutional LSTM extends the fully connected LSTM to have convolutional structures in both the input-to-state and state-to-state transitions. The experimental results show that the convolutional LSTM outperforms the fully connected LSTM for precipitation prediction. In [32], the convolutional architecture with piecewise max pooling utilizes all local features to perform relation extraction globally. The experimental results show that the proposed method achieves better performance than the baseline methods for relation extraction. In [12], based on the single link, the global-level representation was proposed to directly capture the relationship in single link. In this paper, we use the global-level representation to integrate the constraints in multiple links.

2.3. Attention Mechanism

Attention mechanism has recently succeeded in many tasks including image classification [13], neural machine translation [14], multimedia recommendation [15], because it may focus on the effective parts of features when different aspects are considered. The attention mechanism is usually implemented by using a neural network based on the corresponding task. In [13], attention mechanism was proposed to address the problem of enormous computation cost in CNN. In [14], attentional mechanism selectively focuses on the effective parts of input sentences during translating to improve the accuracy of machine translation. In [15], the two-layer attention mechanism is proposed to adaptively extract the implicit feedback. In this paper, the attention mechanism is proposed to use interval times as aspect for traffic prediction.
In [35], the convolutional block attention module was proposed for object detection. Given an intermediate feature map, the proposed module sequentially infers attention maps along two separate dimensions, channel and spatial, then the attention maps are multiplied to the input feature map for adaptive feature refinement. The experimental results show that the proposed method outperforms all the baselines. The attention mechanism in our work is similar to the channel attention in [35].

3. Methodology

3.1. Neural Network Architecture

In the paper, the traffic prediction task is described as follows. Given the traffic data on some links, the traffic time within time interval ( t + δ ) on the selected link will be predicted. Where T is the current interval time and δ is some span of interval time.
The architecture of the proposed method for traffic prediction is illustrated in Figure 2. The input sequences including attribute travel speed and attribute interval time will be fed into the architecture one by one. The predicted travel-time will be computed based on the predicted travel speed and the length of the selected link. Based on the input sequences from multiple links, the global level representation provide the constraints in the task domain for our proposed method [12]. The convolution extracts features from the global-level representation of multiple links for traffic prediction. The attention mechanism is a glimpse network and is implemented by a two-layers neural network. The attention mechanism adaptively select a sequence of regions and only process the selected regions when interval time aspects are considered. The convolutional results will be flattened and fed to a single layer neural network to generate the predicted value.
Interval time is 15 min in the given database, thus there are 96 interval times per day. 96 is the length of interval times embeddings that is a vector and will be learned during model training. Each item of the interval times embedding represent one interval time.
The architecture of our proposed method includes the interval times embeddings, the input sequences component, the global-level representation component, the convolution component, the attention mechanism component and the out predictor.

3.2. Input Sequence

Selecting attributes from the given dataset is the important step of data preparation procedure of data process [36]. The selected attributes in the CNN based methods are usually the same as the attributes to be predicted [8,9,10]. While the success of these CNN-based methods, the attribute interval time is not selected as the attribute of these CNN-based methods that represents the temporal-level structure of the traffic data. In our work, the attribute interval time is selected.
Let current interval time be t and the length of the input sequence be 7, then the input sequence in Figure 1 is described in Equations (1) and (2). Equation (1) denotes the data from the link to be predicted. Equations (2) denotes the data from other links including on-ramps, off-ramps or other adjacent links, i is the index of the multiple links, where x is C F S F . S F is the attribute interval time regarding the temporal-level structure and C F is the attribute travel-time regarding the content. Where the subscripts { t 6 , t 5 , t 4 , t 3 , t 2 , t 1 , t } are interval times. The predicted value is at interval time t + δ and δ denotes time lag that is the interval time of the predicted value and current interval time.
s : x t 6 , x t 5 , x t 4 , x t 3 , x t 2 , x t 1 , x t
s i : x t 6 i , x t 5 i , x t 4 i , x t 3 i , x t 2 i , x t 1 i , x t i

3.3. Convolution

The learning ability of CNN may be enhanced by providing constraints in the task domain [34], in which the constraints in the task domain include the local receptive fields and the shared weights. In our work, the global-level representation was proposed that each x in the input sequence convolutes x t 1 in the input sequence, respectively [12]. The global-level representation is the substitute for the local receptive fields of the convolution component to apply the constraints in traffic prediction task. The constraints is that each historical data point in the input sequence directly effect the predicted value [12].
The convolution is a linear transformation described in Equation 3. Where X ( i , j ) = ( x t i , x t ) is one global-level representation. i ( 1 , , L ) and L is the length of the input sequence. t is the length of training dataset or testing dataset. Filter W k , k ( 1 , , n 1 ) is the shared parameter matrix and hyperparameter n 1 is the number of the filters. Matrix C R n 1 × t denotes the convolutional result. C ( i ; j ) denotes the convolutional result of global-level representation X ( i , j ) . In Equation (3), weight matrix W k is shared across all times.
C ( i , j ) = W k X ( i , j ) 1 < i L 0 < j t k ( 1 , , n 1 )

3.4. Attention Mechanism

Attention mechanism is implemented by constructing a glimpse network according to the corresponding tasks and has recently demonstrated success [13,14,15,35]. The attention mechanism in literature [13] is over the input images and adaptively select a sequence of regions of the input images and only process the selected regions. In our work, the attention mechanism is over the convolutional results and adaptively select a sequence of regions of the convolutional results and only process the selected regions. The attention mechanism of the proposed model is illustrated in Figure 3 and its transition functions are described in Equations (4)–(6). Where matrix H denotes the convolutional results illustrated in Figure 2. v a is the embedding of the interval time. Vector α denotes the attention weights to matrix H. Matrix r denotes the element-wise multiplication between matrix H and α . Specifically, based on an interval time aspect, attention mechanism allows the proposed model to attend over matrix H based on an interval time and may effectively extracts the features in matrix H.
M = t a n h W h H W v v a e n
α = w T M
r = H α T

3.5. Output

The predictor is a single layer fully-connected neural network and is described in Equation (7). Where matrix W 2 R l × n 1 denotes the parameter of the predictor. The predictor receives the output of attention mechanism r to compute the predicted travel-time h a t y , which is the travel-time across the given link within time interval T + δ .
y ^ = W 2 r

3.6. Model Training

The parameters of the proposed method is θ : { W k , W 2 , W h , W v , w , V a , n 1 , l } , k { 1 , , n 1 } . Where { W k , W 2 , W h , W v , w , V a } are randomly initialized and updated during the training of the proposed method. n 1 is a hyper-parameter and is the number of the kernels of the convolution compnent. l is the length of the input sequence. The proposed method is trained by using the back-propagation method. The cost function of the proposed method is described in Equations (4)–(6). Where S = { ( s i , y i ) } denotes the training examples. { s i } denotes the input sequences. y i denotes the travel-time within time interval T + δ . y i ^ denotes the predicted travel-time within time interval T + δ .
The mini-batch gradient descent method is used to compute the cost function and the partial derivative of parameters θ . AdaGrad is used to update parameters θ . AdaGrad is described in Equation (6), where ε is a smoothing term that avoids division by zero (usually on the order of 1 × 10 8 ). AdaGrad performs the larger updates for infrequent parameters rather than frequent parameters. Thus, the need to manually tune the learning rate of gradient descent methods is eliminated and the robustness of gradient descent methods is greatly improved [37]. Theano is used to develop the code of the proposed method. The proposed method is trained on a 4-G GPU computer with CUDA [38] in which CNMeM is enabled with an initial size by using 80.0% of the memory.
We choose hyperparameters n 1 based on validation set, adjust parameter matrices { W 1 , W 2 , W h , W v , w , V a } based on training set and evaluate the accuracy of the proposed method based on testing set.
J ( θ ) = S 1 | S | y ^ θ ( s i ) y i 2
G t , i = t θ i J ( θ ) 2
θ t + 1 , i = θ t , i η G t , i + ε · θ i J ( θ )

3.7. Dataset and Task Definition

We evaluate the performance of the proposed method based on the dataset provided by Highways England [39]. Highways England have maintained and improved the motorways and major A roads in the England. The dataset for our experiment is described in Table 2, where AverageJT is the travel-time that is taken by vehicles to across LinkRef within TimePeriod. TimePeriod denotes time interval that vary from 0 to 95 and is 15 min.
The database of the links spans from 1 March 2015 to 31 March 2015. We divide the dataset into three parts including the training set, the validation set and the testing set, which contain data from 1 March to 27 March (approximately 87.1%), March 28 (approximately 3.2%) and 29 March to 31 March (approximately 9.7%), respectively.
Input Sequence and Its Output Value: Let we predict the travel-time across AL282 within time interval t + δ . The input sequence is described in Equation (11), where δ is a time lag and C F is the travel-time attribute.
S : x t 6 AL 282 x t 5 AL 282 x t 4 AL 282 x t 3 AL 282 x t 2 AL 282 x t 1 AL 282 x t AL 282 x t 6 AL 283 x t 5 AL 283 x t 4 AL 283 x t 3 AL 283 x t 2 AL 283 x t 1 AL 283 x t AL 283 x t 6 AL 292 x t 5 AL 292 x t 4 AL 292 x t 3 AL 292 x t 2 AL 292 x t 1 AL 292 x t AL 292 x t 6 AL 286 x t 5 AL 286 x t 4 AL 286 x t 3 AL 286 x t 2 AL 286 x t 1 AL 286 x t AL 286 x t 6 AL 2274 x t 5 AL 2274 x t 4 AL 2274 x t 3 AL 2274 x t 2 AL 2274 x t 1 AL 2274 x t AL 2274
y ^ = x t + δ AL 282 . C F

4. Experimental Setting

Normalization: To avoid calculation overflow during the training process and better capture the nonlinear relationships, C F and S F are rescaled to ( 1 , 1) because ( 1 , 1) is the range of the activation function in the proposed method.

4.1. Evaluation Metrics

To evaluate the performance of the predictive methods, we adopt three performance indices including the mean absolute error (MAE), the mean absolute percentage error (MAPE) and the RMS error (RMSE). These indices are stated as Equations (13)–(15), respectively. Where S denotes the input sequences. i is the index input sequence in S. y i denotes the observed value and y i ^ denotes the output value.
MAE = 1 | S | i = 1 | S | | y i ^ y i |
MAPE = 1 | S | i = 1 | S | | y i ^ y i | | y i |
RMSE = [ 1 | S | i = 1 | S | ( y i ^ y i ) 2 ] 1 2

4.2. Hyperparameters

In this section, we experimentally determine hyperparameter { n 1 , l } based on the validation set. l, the length of an input sequence and n 1 , the number of the kernels of the convolution compnent. The hyperparameter values is determined in a certain range as follows: l in { 2 , , 10 } and n 1 in { 1 , 5 , 10 , 20 , , 60 } . Figure 4 describes the variety of the MAPE with parameters l, n 1 increasing. The optimal values of l and n 1 are 3 and 30, respectively. The optimal learning rate of the parameters for AdaGrad, η , is 0.1. Table 3 describes the hyperparameter values that are used in our experiments.

4.3. Baseline Methods

Eight methods are selected as the baseline methods to evaluate the performance of our proposed method. These methods include SARIMA, LR, k-NNR, SVR, SSNN, LSTM NN, CNN and AEs. In our experiments, we predict the travel-time within time interval t + δ on link AL282 where lag δ is one time interval.
The LR method in our experiments is based on a publication by Rice and Zwet [20] in which the predicted travel-time is the linear regression of the travel-time within time interval t and the mean travel-time within historical time interval t + δ .
The SARIMA method in our experiments is based on the contributions in the literature [18] in which the input sequences are considered as a SARIMA ( 1 , 0 , 1 ) ( 0 , 1 , 1 ) S process. We let hyperparameter S be 96 because there are 96 time intervals per day in the our dataset.
The Kalman filter method in our experiment is based on the contributions in literature [19]. The traffic values observed on the previous day was used in place of the observation for correcting the apriori estimate.
In the k-NNR method in our experiments, the predicted value is a weighted average value of the k-NNs in which the time intervals are before time interval t + δ [22]. The weights that are assigned to the k-NNs are usually accomplished by using a distance scheme or a uniform scheme. The distance scheme assigns different weights, usually the inverse of their distance to the predicted data point, to the contributions of the k-NNs. The uniform scheme assigns the same weights to the contributions of the k-NNs.
The SVR method with an RBF kernel is based on the contributions in literature [25]. The SVR method in our experiments uses three types of kernels, namely, a linear kernel (linear), a polynomial kernel (poly), and an RBF kernel.
The SSNN method is based on the contributions in literature [4] in which the connection between the internal states and the context layers is fixed at 1.0 and the values of the nodes in the context layer are initially set to 0.5. The connection and the values are not subject to adjustment in the training process. In our experiments, we let the number of nodes in the hidden layer and the output layer be 10. We let the length of the input sequence and the unfolded size of SSNN be 12.
The LSTM NN is based on a publication in literature [5]. In our experiments, we let the unfolded size of the LSTM NN be the same as literature. The AEs method in our experiments is based on a publication in literature [6]. The CNN1 method in our experiments is based on a publication in literature [8]. The CNN2 method in our experiments is based on the publication in literature [12].
In addition, we parameterize the baseline methods as described in the original literatures, in which the baseline methods have been tuned for other datasets. This hints that they perhaps perform considerably better on this dataset if they also gets tuned for it. For future work, we will turn the parameters of the baseline methods to perform a further comparison.

5. Experimental Results

5.1. Exploring the Spatiotemporal Relationship

The CNN-based method is effective because there is exist spatiotemporal relationship in multiple links. We demonstrate the relationship by using Pearson correlation in Equation (16), where X, Y denote two random variables with the same number of observations. In literature [9], the Pearson correlation is also used to explore the spatiotemporal relationship of the variables in short-term passenger demand prediction. We calculate the Pearson correlations between the predicted travel-time at interval time t in link i and the history travel-time at time interval t k in link j. Where (i = AL282, j {AL286, AL292, AL2274, AL282, AL283}) or (i = AL281, j {AL288, AL291, AL281, AL284}), k { 1 , 2 , 3 , 4 , 5 , 6 , 7 } .
C o r r ( X , Y ) = E ( X Y ) E ( X ) E ( Y ) E ( X 2 ) E 2 ( X ) ( Y 2 ) E 2 ( Y )
Figure 5 shows the relationship in multiple links that is correlations between the dependent variable (the to-be-predicted travel-time at interval time t on Link i) and the explanatory variables (the observed travel-time at interval time t k on Link j).
It can show that there are strong temporal correlations among dependent variable and the explanatory variables, which drop gradually with the increase of time intervals. On the other hand, the variables with shorter spatial distance have strong correlations, but the variables with longer spatial distance are also correlated with the to-be-predicted travel-time to some extent.

5.2. Performance with Time Variation

We study the prediction performances of the proposed method during 3 day under different traffic conditions where the time lag is one time interval. Figure 6a illustrates the comparison between the observed values and the predicted values on link AL282. Figure 6b illustrates the comparison between the observed values and the predicted values on link AL281. The comparison can be well visualized when zooming into a finer grade. Figure 6 shows that the observed values and the predicted values are in good agreement and the proposed method is effective. Figure 6 shows that the proposed method may capture the sudden changes in travel-time and has a tendency to underestimate the future traffic condition. Specially, the proposed method may capture the sudden regime changes from free flow to congestion and then the recovery regime.
Figure 7 illustrates that the absolute values of the residuals (red circles) against the observed data (blue line) during 3 days on link AL281. We can see that the performance of the proposed method is not uniform through days. Several large errors are observed at when regime changes from free flow to congestion and starts to recover back to free flow. Highest residuals are observe at when travel-time regime changes from one to another.
Heat hots Figure 8 is another visual way to interpret the results of prediction where the deeper color implies the longer travel-time. The congestion propagation on traffic network led traffic prediction to complex. For example, we consider a stretch of highway and assume a bottleneck, it is expected that the end of the queue will move from the bottleneck downstream. Both the head and tail of the bottleneck sometimes move downstream together. Figure 8 shows that the proposed method properly may capture both forward and backward shock wave propagation on link AL281 during 3 days.
Figure 9 plots the cumulative distribution functions of the absolute prediction errors on links AL281 and AL282 to demonstrate the statistical property of the proposed method, CNN and AES, respectively. The time interval is 15 min. Figure 9 shows that the prediction error of the proposed method is smaller than CNN and AEs on link AL281 and link AL282. In summary, the experimental results testified the effectiveness of the proposed method for travel-time prediction.

5.3. Overall Performance

To compare the proposed method with the baseline methods, we perform the experiments that predict the travel-time across link AL282 within time interval T + 1 based on multiple links AL286, AL292, AL2274 AL282 and AL283. We perform the experiments that predict the travel-time across link AL281 within time interval T + 1 based on multiple links AL288, AL291, AL281, AL284 and AL279. The length of link AL282 is approximately 15.64 km and the length of link AL281 is approximately 16.00 km. The multiple links is illustrated in Figure 3.
Table 4 shows the experimental results across link AL282. The experimental results show that the proposed method achieves better accuracy than the baseline methods. Table 5 shows the experimental results across link AL281. The experimental results shows that the proposed method achieves better accuracy that the baseline methods.
Discussion of Results: Prediction errors may usually be decomposed into two subcomponents including bias error attributed to erroneous assumptions and variance error attributed to the sensitivity to small fluctuations. High bias denotes that the methods miss some relevant relations between the features and the predictive outputs (under fitting). Low bias but high variance denotes that the random noises in the training dataset are transformed by the methods rather than the intended outputs (over fitting) [40]. Based on the multiple links illustrated in Figure 3, Table 4 and Table 5 show that the proposed method outperforms all baseline methods in terms of bias error and the variance of the proposed method has the same order of magnitude as the NN-based methods. Thus, the proposed method has no under-fitting or over-fitting problems and is an effective approach for travel-time prediction.
Time Costs: Figure 10 shows the execution times spent by the proposed method and the baseline methods based on link AL282. According to the time costs, these methods are classified into three groups including SARIMA method, NN-based methods and the remaining methods. SARIMA method takes the most time to execute, the NN-based methods take the moderate time to execute and the remaining methods take the least time to execute. The proposed method is within a reasonable boundary of time consumption.

5.4. Performance for the Individual Links

To validate the performance of the proposed method, we perform the experiments on every road, which are AL344, AL3276, AL444, AL2878, AL2869, AL2871, AL2861, AL2853, AL2850, AL2852B, AL2852A, AL286, AL292, AL283, AL278, AL270, AL265, AL261A, AL258A, AL256, AL2282, AL248, AL241, AL242, AL236, AL2292, AL2295, AL237, AL243, AL240, AL254, AL257A, AL267, AL272, AL279, AL284, AL291, AL288, AL2851A, AL2851B, AL2849, AL298, AL2862, AL343, AL2872, AL2870 and AL2879. There are 47 links. For each link, we cluster five adjacent links as multiple links for training and testing. we perform the experiments to predict travel-time across these links. The baseline methods include LR, LSTM, SSNN, AES, CNN1 [8] and CNN2 [12]. The experiment results are described in Table 6 and in Figure 11. Where the interval time is 15 min.
Table 6 and Figure 11 shows the MAPE of these links. Table 6 and Figure 11 shows that the proposed method outperforms the baseline methods on these links. Table 6 and Figure 11 indicates that the prediction errors of different links vary greatly. The predictability of links #10, #12, #13, #14, #16, #34, #37 and #38 ( AL2852, AL286, AL292, AL283, AL270, AL272 and AL291 ) are much better, the predictability of links #17, #30, #34 and #44 (AL265, AL240, AL272 and AL343) are much poorer.

6. Conclusions and Future Work

In this paper, we propose the convolution component-based method with attention mechanism for traffic prediction. The global level representation is used based on multiple links. The time intervals is considered as the aspect of attention mechanism. Based on the dataset provided by Highways England, the experimental results show that the proposed method can improve the predictive accuracy than the baseline methods.
The proposed method still exist a weakness that it have the bigger absolute prediction errors at some time intervals. Improving the weakness may be conducted as a future work. Another future work is providing a better substitute for the predicted travel-time in the input sequences.

Author Contributions

Conceptualization, X.R., Z.S. and Y.F.; Funding acquisition, Z.S.; Investigation, X.R.; Methodology, X.R., Z.S. and Y.F.; Project administration, Z.S.; Supervision, Z.S. and C.L.; Validation, X.R.; Writing—original draft, X.R.; and Writing—review and editing, Y.F.

Funding

This research was funded in part by the National Key R & D Program of China (grant numbers 2018YFB0803400 and 2018YFB0505302) and the National Science Foundation of China (grant number 61832012).

Acknowledgments

We are thankful to Jicheng Zhou for the laboratory equipment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moria, U.; Mendiburub, A.; Álvarezc, M.; Lozanoa, J.A. A review of travel-time estimation and forecasting for advanced traveller information systems. Transportmetrica 2015, 11, 119–157. [Google Scholar] [CrossRef]
  2. Van Lint, J.W.C. Online learning solutions for freeway travel-time prediction. IEEE Trans. Intell. Transp. Syst. 2008, 9, 38–47. [Google Scholar] [CrossRef]
  3. Liu, H.; Zuylen, H.J.V.; Lint, H.V.; Salomons, M.; Liu, H. Predicting urban arterial travel-time with state-space neural networks and kalman filters. Transp. Res. Rec. J. Transp. Res. Board 2006, 1968. [Google Scholar] [CrossRef]
  4. Lint, H. Reliable Travel-Time Prediction For Freeways. Ph.D. Thesis, Delft University of Technology, Delft, The Nederland, 3 May 2004. [Google Scholar]
  5. Ma, X.; Tao, Z.; Wang, Y.; Yu, H.; Wang, Y. Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp. Res. Part C 2015, 54, 187–197. [Google Scholar] [CrossRef]
  6. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic flow prediction with big data: A deep learning approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar] [CrossRef]
  7. Polson, N.G.; Sokolov, V.O. Deep learning for short-term traffic flow prediction. Transp. Res. Part C Emerg. Technol. 2017, 79, 1–17. [Google Scholar] [CrossRef]
  8. Ma, X.; Dai, Z.; He, Z.; Ma, J.; Wang, Y.; Wang, Y. Learning traffic as images: A deep convolutional neural network for large-scale transportation network speed prediction. Sensors 2017, 17, 818. [Google Scholar] [CrossRef] [PubMed]
  9. Ke, J.; Zheng, H.; Yang, H.; Chen, X. Short-term forecasting of passenger demand under on-demand ride services: A spatio-temporal deep learning approach. Transp. Res. Part C Emerg. Technol. 2017, 85, 591–608. [Google Scholar] [CrossRef]
  10. Wang, J.; Gu, Q.; Wu, J.; Liu, G.; Xiong, Z. Traffic speed prediction and congestion source exploration: A deep learning method. In Proceedings of the IEEE 16th International Conference on Data Mining, Barcelona, Spain, 12–15 December 2017; pp. 499–508. [Google Scholar]
  11. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  12. Ran, X.; Shan, Z.; Fang, Y.; Lin, C. Travel-time prediction by providing constraints on a convolutional neural network. IEEE Access 2018, 6. [Google Scholar] [CrossRef]
  13. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst. 2014, 3, 2204–2212. [Google Scholar]
  14. Luong, M.T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
  15. Chen, J.; Zhang, H.; He, X.; Liu, W.; Liu, W.; Chua, T.S. Attentive Collaborative Filtering: Multimedia Recommendation with Item- and Component-Level Attention. In Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 335–344. [Google Scholar]
  16. Ahmed, M.S.; Cook, A.R. Analysis of Freeway Traffic Time-Series Data by Using Box-Jenkins Techniques. Transp. Res. Rec. J. Transp. Res. Board 1979, 773, 1–9. [Google Scholar]
  17. Hamed, M.M.; Al-Masaeid, H.R.; Said, Z.M.B. Short-term prediction of traffic volume in urban arterials. J. Transp. Eng. 1995, 121, 249–254. [Google Scholar] [CrossRef]
  18. Williams, B.M.; Hoel, L.A. Modeling and forecasting vehicular traffic flow as a seasonal arima process: Theoretical basis and empirical results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef]
  19. Kumar, S.V. Traffic Flow Prediction using Kalman Filtering. Procedia Eng. 2017, 187, 582–587. [Google Scholar] [CrossRef]
  20. Rice, J.; Van Zwet, E. A simple and effective method for predicting travel times on freeways. IEEE Trans. Intell. Transp. Syst. 2004, 5, 200–207. [Google Scholar] [CrossRef]
  21. Davis, G.A.; Nihan, N.L. Nonparametric regression and short-term freeway traffic forecasting. J. Transp. Eng. 1991, 117, 178–188. [Google Scholar] [CrossRef]
  22. Smith, B.L.; Williams, B.M.; Oswald, R.K. Comparison of parametric and nonparametric models for traffic flow forecasting. Transp. Res. Part C Emerg. Technol. 2002, 10, 303–321. [Google Scholar] [CrossRef]
  23. Chang, H.; Lee, Y.; Yoon, B.; Baek, S. Dynamic near-term traffic flow prediction: Systemoriented approach based on past experiences. IET Intell. Transp. Syst. 2012, 6, 292–305. [Google Scholar] [CrossRef]
  24. Drucker, H.; Burges, C.J.C.; Kaufman, L.; Smola, A.J.; Vapnik, V. Support vector regression machines. Adv. Neural Inf. Process. Syst. 1996, 28, 779–784. [Google Scholar]
  25. Wu, C.H.; Ho, J.M.; Lee, D.T. Travel-time prediction with support vector regression. IEEE Trans. Intell. Transp. Syst. 2004, 5, 276–281. [Google Scholar] [CrossRef]
  26. Castro-Neto, M.; Jeong, Y.S.; Jeong, M.K.; Han, L.D. Online-svr for short-term traffic flow prediction under typical and atypical traffic conditions. Expert Syst. Appl. Int. J. 2009, 36, 6164–6173. [Google Scholar] [CrossRef]
  27. Su, H.; Zhang, L.; Yu, S. Short-term traffic flow prediction based on incremental support vector regression. In Proceedings of the 3rd International Conference on Natural Computation, Haikou, China, 24–27 August 2007; pp. 640–645. [Google Scholar]
  28. Lint, H.V.; Hoogendoorn, S.P.; Zuylen, H.J.V. State space neural networks for freeway travel-time prediction. In Proceedings of the International Conference on Artificial Neural Networks (ICANN 2002), Madrid, Spain, 28–30 August 2002; pp. 1043–1048. [Google Scholar]
  29. Elman, J. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  30. Van Lint, H.; Van Hinsbergen, C. Short-term traffic and travel-time prediction models. Transp. Res. Circ. 2012, E-C168, 22–41. [Google Scholar]
  31. Zeng, X.; Zhang, Y. Development of recurrent neural network considering temporal-spatial input dynamics for freeway travel-time modeling. Comput.-Aided Civ. Infrastruct. Eng. 2013, 28, 359–371. [Google Scholar] [CrossRef]
  32. Zeng, D.; Liu, K.; Chen, Y.; Zhao, J. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1753–1762. [Google Scholar]
  33. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional lstm network: A machine learning approach for precipitation nowcasting. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 9199, pp. 802–810. [Google Scholar]
  34. Lecun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  35. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  36. Wirth, R.; Hipp, J. CRISP-DM: Towards a standard process model for data mining. In Proceedings of the 4th International Conference on the Practical Application of Knowledge Discovery and Data Mining, Manchester, UK, 11–13 April 2000. [Google Scholar]
  37. Dean, J.; Corrado, G.S.; Monga, R.; Chen, K.; Devin, M.; Le, Q.V.; Mao, M.Z.; Ranzato, M.; Senior, A.; Tucker, P.; et al. Large scale distributed deep networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1223–1231. [Google Scholar]
  38. Vanhoucke, V.; Mao, M.Z. Improving the speed of neural networks on CPUs. Deep Learning & Unsupervised Feature Learning Workshop Nips. Available online: http://www.andrewsenior.com/papers/VanhouckeNIPS11.pdf (accessed on 1 May 2019).
  39. Highways England. Highways Agency Network Journey Time and Traffic Flow Data; Highways England: Guildford, UK, 2018. [Google Scholar]
  40. Fortmannroe, S. Understanding the Bias-Variance Tradeoff. Available online: http://scott.fortmann-roe.com/docs/BiasVariance.html (accessed on 1 May 2019).
Figure 1. A section of road network that contains multiple links.
Figure 1. A section of road network that contains multiple links.
Sensors 19 02063 g001
Figure 2. Architecture of the convolutional neural networks wich attention mechanism for travel-time prediction.
Figure 2. Architecture of the convolutional neural networks wich attention mechanism for travel-time prediction.
Sensors 19 02063 g002
Figure 3. Architecture of attentive neural network.
Figure 3. Architecture of attentive neural network.
Sensors 19 02063 g003
Figure 4. Variety of the MAP with parameters l and n 1 in the proposed method. (a) Size of an input sequence. (b) Number of the kernels of the convolution compnent.
Figure 4. Variety of the MAP with parameters l and n 1 in the proposed method. (a) Size of an input sequence. (b) Number of the kernels of the convolution compnent.
Sensors 19 02063 g004
Figure 5. Correlations in multiple links. (a) Time intervals on Link AL281. (b) Time intervals on Link AL282.
Figure 5. Correlations in multiple links. (a) Time intervals on Link AL281. (b) Time intervals on Link AL282.
Sensors 19 02063 g005
Figure 6. Effectiveness of the prediction under the different traffic conditions. The horizontal axis is time interval and the vertical axis is travel-time. (a) Comparison between the observed values and the predicted values on Link AL282. (b) Comparison between the observed values and the predicted values on Link AL281.
Figure 6. Effectiveness of the prediction under the different traffic conditions. The horizontal axis is time interval and the vertical axis is travel-time. (a) Comparison between the observed values and the predicted values on Link AL282. (b) Comparison between the observed values and the predicted values on Link AL281.
Sensors 19 02063 g006
Figure 7. Absolute values of the residuals against the observed data during 3 days on link AL281. The horizontal axis is time intervals and the vertical axis is travel-times.
Figure 7. Absolute values of the residuals against the observed data during 3 days on link AL281. The horizontal axis is time intervals and the vertical axis is travel-times.
Sensors 19 02063 g007
Figure 8. Heat plots of travel-time prediction during 3 days. The horizontal axis is time intervals.
Figure 8. Heat plots of travel-time prediction during 3 days. The horizontal axis is time intervals.
Sensors 19 02063 g008
Figure 9. Cumulative distribution of the absolute predictive errors on link AL281 and link AL282.
Figure 9. Cumulative distribution of the absolute predictive errors on link AL281 and link AL282.
Sensors 19 02063 g009
Figure 10. Time costs of the proposed method and the baseline methods.
Figure 10. Time costs of the proposed method and the baseline methods.
Sensors 19 02063 g010
Figure 11. Experimental results of the proposed method and the baseline methods on individual links.
Figure 11. Experimental results of the proposed method and the baseline methods on individual links.
Sensors 19 02063 g011
Table 1. Advantages and disadvantages of individual deep learning methods.
Table 1. Advantages and disadvantages of individual deep learning methods.
MethodsAdvantages and Disadvantages
SSNNUse shared internal state (memory) to process sequences of inputs and exists the issue of back-propagated error decay through memory blocks [4].
LSTM NNCan overcome the issue of back-propagated error decays through memory blocks [5].
AESSpatial and temporal correlations are inherently considered [6].
CNNTemporal correlations are inherently considered [12,32,33]. Based on the topology of the input data, three architectural ideas are proposed including local receptive fields, shared weights and spatial or temporal subsampling [34].
Table 2. Description of the dataset provided by Highways England.
Table 2. Description of the dataset provided by Highways England.
LinkRefDateTimePeriod (0–95)AverageJTLinkLength (km)
AL2822015-03-010642.5615.64
AL2822015-03-011603.7115.64
AL2922015-03-010126.033.92
AL2922015-03-011130.523.92
AL22742015-03-010337.409.72
AL22742015-03-011337.409.72
AL2862015-03-010271.837.98
AL2862015-03-011272.697.98
Table 3. Hyper parameters value used in our experiments.
Table 3. Hyper parameters value used in our experiments.
Hyperparameterl n 1 η
Value3300.1
Table 4. Experimental results of the proposed method and the baseline methods on link AL282. The best results in the table appear in bold.
Table 4. Experimental results of the proposed method and the baseline methods on link AL282. The best results in the table appear in bold.
Models 15-min on Training Set 15-min on Test SetVariance
RMSEMAEMAPERMSEMAEMAPE
k-NNRuniform54.5445.7455.90 53.3615.9116.550.65
distance33.5124.3083.34 53.3615.9116.553.21
SARIMA 193.55413.12030.03 55.8206.6588.5321.5
SVRlinear54.6126.0406.71 49.6896.0467.050.34
rbf57.0776.4417.69 51.3866.3107.740.05
poly58.1366.8949.29 66.5067.1359.470.18
LR 54.8536.1956.04 52.5125.7867.081.04
SSNN 52.9015.7065.81 47.8495.7946.260.45
LSTM NN 52.2705.7285.89 48.1965.8116.300.41
AEs 60.0416.1176.70 48.3266.0286.810.11
CNN1 64.0407.0819.06 95.9157.6539.820.76
CNN2 51.7725.7756.02 53.8295.8046.320.30
The Proposed Method49.7945.5545.79 52.6445.6756.200.45
Table 5. Experimental results of the proposed method and the baseline methods on link AL281. The best results in the table appear in bold.
Table 5. Experimental results of the proposed method and the baseline methods on link AL281. The best results in the table appear in bold.
Models 15-min on Training Set 15-min on Test SetVariance
RMSEMAEMAPERMSEMAEMAPE
k-NNRuniform20.1733.1721.88 42.9415.3075.453.57
distance44.6635.2355.14 42.9415.3075.450.31
SARIMA 162.66912.24226.63 52.8266.3468.0218.61
SVRlinear41.9945.3465.26 40.2535.3495.700.44
rbf45.0945.7896.22 45.0565.8336.850.63
poly48.5786.0716.84 49.2006.1287.590.75
LR 38.7895.2195.30 48.1415.7466.050.75
SSNN 41.2905.2814.98 38.7375.1825.210.23
LSTM NN 41.0975.1964.90 37.8455.1125.140.24
AEs 47.4045.5635.58 41.6385.6156.130.55
CNN1 51.03475.906.29 45.7725.9686.870.58
CNN2 40.3675.1664.87 42.345.1545.160.29
The Proposed Method39.2785.0814.79 41.195.0245.020.23
Table 6. Experimental results of the proposed method and the baseline methods on individual links in terms of MAPE. The best results in the table appear in bold.
Table 6. Experimental results of the proposed method and the baseline methods on individual links in terms of MAPE. The best results in the table appear in bold.
LinkIDKalman FilterLRCNN2LSTMAESCNN 1SSNNThe Proposed Method
AL34413.895.836.095.6810.9818.426.035.46
AL327610.875.695.635.376.9313.275.535.17
AL4448.486.025.986.168.3827.286.495.97
AL287816.0310.347.968.4114.112.838.937.66
AL286913.316.856.897.5612.724.247.886.84
AL287113.398.586.646.7910.6411.37.486.03
AL286110.7215.837.827.6110.6614.417.917.46
AL28538.435.084.735.327.315.035.174.56
AL28508.066.126.466.428.839.956.546.13
AL2852B5.312.021.821.82.114.781.941.75
AL2852A6.415.445.325.385.789.285.415.27
AL2865.171.291.251.261.925.051.521.21
AL2924.561.421.292.431.934.822.241.26
AL2835.092.811.941.872.594.781.961.9
AL27813.3411.0810.8611.8213.0415.1612.1210.81
AL2705.072.751.831.752.45.072.021.78
AL2659.956.376.116.248.5210.176.375.9
AL261A10.056.886.776.618.8710.866.586.54
AL258A6.645.335.725.676.768.635.845.46
AL2569.859.567.157.09.299.666.926.92
AL22828.056.276.226.17.779.546.376.17
AL2488.445.395.655.647.119.315.65.56
AL2418.896.416.196.086.877.846.555.96
AL2427.599.356.956.847.4210.146.956.77
AL2368.577.667.167.368.110.717.747.05
AL22929.367.987.717.929.358.547.757.57
AL22957.305.465.15.386.376.835.314.65
AL2376.846.296.326.417.16.846.66.08
AL2437.9410.45.556.266.6812.465.775.5
AL2407.836.746.096.036.868.786.495.97
AL2549.858.537.727.579.349.937.937.48
AL257A7.076.56.26.087.167.636.526.08
AL2677.506.425.025.366.737.555.154.96
AL2723.951.951.411.721.723.331.641.33
AL27919.2616.7514.8917.1219.7926.3817.014.04
AL2845.886.296.176.16.928.866.546.06
AL2914.862.1.261.292.013.91.491.20
AL2886.242.431.711.743.015.021.831.64
AL2851A6.985.835.85.696.086.595.815.76
AL2851B9.496.115.936.258.4110.556.565.88
AL28498.758.65.626.27.7812.165.95.38
AL29810.566.586.075.648.1328.135.735.39
AL28627.276.436.196.147.247.756.845.7
AL34311.269.215.26.4411.159.316.115.03
AL28725.426.684.885.026.069.055.664.72
AL28706.585.95.825.957.059.966.175.75
AL287914.137.126.766.6210.7226.16.876.73
Mean8.736.615.705.887.5911.036.0385.50

Share and Cite

MDPI and ACS Style

Ran, X.; Shan, Z.; Fang, Y.; Lin, C. A Convolution Component-Based Method with Attention Mechanism for Travel-Time Prediction. Sensors 2019, 19, 2063. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092063

AMA Style

Ran X, Shan Z, Fang Y, Lin C. A Convolution Component-Based Method with Attention Mechanism for Travel-Time Prediction. Sensors. 2019; 19(9):2063. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092063

Chicago/Turabian Style

Ran, Xiangdong, Zhiguang Shan, Yufei Fang, and Chuang Lin. 2019. "A Convolution Component-Based Method with Attention Mechanism for Travel-Time Prediction" Sensors 19, no. 9: 2063. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092063

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop