Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
A Short Review on the Role of the Metal-Graphene Hybrid Nanostructure in Promoting the Localized Surface Plasmon Resonance Sensor Performance
Previous Article in Journal
A Highlight-Generation Method for Rendering Translucent Objects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An LSTM-Based Method with Attention Mechanism for Travel Time Prediction

1
School of Computer Science and Technology, University of Science and Technology Beijing, Beijing 100083, China
2
Informatization and Industry Development Department, State Information Center, Beijing 100045, China
3
College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China
4
Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China [email protected]
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(4), 861; https://doi.org/10.3390/s19040861
Submission received: 2 January 2019 / Revised: 5 February 2019 / Accepted: 12 February 2019 / Published: 19 February 2019

Abstract

:
Traffic prediction is based on modeling the complex non-linear spatiotemporal traffic dynamics in road network. In recent years, Long Short-Term Memory has been applied to traffic prediction, achieving better performance. The existing Long Short-Term Memory methods for traffic prediction have two drawbacks: they do not use the departure time through the links for traffic prediction, and the way of modeling long-term dependence in time series is not direct in terms of traffic prediction. Attention mechanism is implemented by constructing a neural network according to its task and has recently demonstrated success in a wide range of tasks. In this paper, we propose an Long Short-Term Memory-based method with attention mechanism for travel time prediction. We present the proposed model in a tree structure. The proposed model substitutes a tree structure with attention mechanism for the unfold way of standard Long Short-Term Memory to construct the depth of Long Short-Term Memory and modeling long-term dependence. The attention mechanism is over the output layer of each Long Short-Term Memory unit. The departure time is used as the aspect of the attention mechanism and the attention mechanism integrates departure time into the proposed model. We use AdaGrad method for training the proposed model. Based on the datasets provided by Highways England, the experimental results show that the proposed model can achieve better accuracy than the Long Short-Term Memory and other baseline methods. The case study suggests that the departure time is effectively employed by using attention mechanism.

Graphical Abstract

1. Introduction

Traffic efficiency is important to travellers and is a key indicator of the level of traffic services, especially with the sharp increase in vehicles and the congestion in transfer in road networks. Traffic prediction can help travellers reasonably arrange in road network and improve traffic efficiency. Thus, traffic prediction is still a hot research topic of intelligent transportation systems (ITS). The evaluation indexes of traffic service level (e.g., queue length, severity of incident, traffic volume and average speed) assist travellers in making decisions pre-trip and en-route. Travel time is very intuitive and easily understood by travellers. Thus, travel time is very useful and widely accepted as an evaluation index, thus travel time prediction is considered in this paper.
Travel prediction is a research topic based on modeling the complex non-linear spatiotemporal traffic dynamics in road network [1]. Various traffic prediction methods such as support vector machine regression (SVR), linear regression (LR), k-nearest neighbor regression (k-NNR), autoregressive integrated moving average (ARIMA), and recurrent state-space neural network (SSNN) have been proposed in this area. CNN methods have been successfully applied to traffic prediction and have achieved better performance in recent years. These methods use the implicit correlations between near neighbors by convolution function [2,3,4]. Long Short-Term Memory(LSM NN)-based methods have been successfully applied to traffic prediction and have achieved better performance in recent years [5,6] because they have a better ability to model the traffic dynamics in road network as they can model long-term dependence in time series and extract features from traffic data with recurrent feedback.
The existing LSTM NNs for traffic prediction have two drawbacks: they do not make use of the departure time through the links for traffic prediction, and the way of model long-term dependence in time series is not direct in terms of traffic prediction. Travel time has a correlation with its departure time because traffic dynamics has some periodicity with departure time when traffic flow is not free. Thus, LSTM NN with departure time can provide better performance in predicting travel time than LSTM NN without departure time. The function of modeling long-term dependence in LSTM NNs is usually finished by extracting context information. It is well known that this method is excellent for nature language process because there are contextual structure in nature language. The data from traffic dynamics have no contextual structure. Thus, we assume that LSTM NN is less effective for traffic prediction. LSTM NN may be improved in traffic prediction because each data point in traffic dynamics may directly affect the predicted value.
Attention mechanism has recently demonstrated success in a wide range of tasks and is implemented by constructing a neural network according to the task. The attention mechanism can concentrate on the differences of input features to better extract features when different aspects are considered.
In this paper, we propose a LSTM NN with attention mechanism for travel time prediction (LSAM) that is over the output layers of LSTM NN. In the proposed LSTM NN, we use attention mechanism to construct the depth of LSTM NN and to model long-term dependence. The attention mechanism integrates departure time into the proposed model. The departure time is used as the aspect of the attention mechanism. We use AdaGrad method for optimizing the proposed model. Based on the datasets provided by Highways England, the experimental results show that the proposed model can achieve better accuracy than the LSTM NN and other baseline methods. The case study suggests that the departure time is effectively employed by using attention mechanism.
The contributions of this paper can be summarized as follows:
  • We propose an LSTM-based method with attention mechanism for travel time prediction. In the proposed model, the traditional recurrent way to construct the depth of LSTM NN for modeling long-term dependence is substituted by an attention mechanism that is over the output layers of LSTM. The departure time is used as the aspect of the attention mechanism and is integrated into the proposed model.
  • Experiments were performed based on the dataset provided by Highways England. The experimental results show that the proposed model can achieve better accuracy than the existing LSTM NN and other baseline methods.
  • The case study suggest that the attention mechanism can effectively concentrate on the differences of input features for travel time prediction. The proposed model is feasible and effective.
The remainder of this paper is organized as follows. Section 2 reviews related works on traffic prediction. Section 3 describes the LSTM-based model with attention mechanism for travel time prediction. Section 4 compares Random Walk (RW), Seasonal Autoregressive Integrated Moving Average (SARIMA), Linear Regression (LR), k-Nearest Neighbor Regression (k-NNR), Support Vector Machine Regression (SVR), Recurrent State-space Neural Network (SSNN) and LSTM NN and discusses the experimental results. Finally the paper discusses the conclusion and future work.

2. Related Work

2.1. Data-Driven Models for Traffic Prediction

Given that enough data are available, data-driven models are more appropriate for traffic prediction, specifically neural network approaches [1]. The function ( f ) of data-driven approaches that relates the explicative variables with the target variable is usually determined by using statistical inference and machine learning techniques. In this paper, these traffic prediction methods fall into two categories: the methods considering departure time and the methods not considering departure time. The methods considering departure time are SARIMA and the LR. The methods that do not consider departure time are k-NNR, SVR and neural networks (NNs).
Seasonal Autoregressive Integrated Moving Average is a time-series analysis method for traffic prediction. ARMA is a combination of Autoregressive (AR) model and Moving Average (MA) model. ARMA is usually used for traffic prediction under stationary traffic dynamics where the mean, variance and auto-correlation are unchanged. ARIMA is a generalization of ARMA. When the traffic dynamic is non-stationarity, the initial differencing step of ARIMA is used one or more times to eliminate the non-stationarity.
Ahmed and Cook [7] proposed using ARIMA to predict short-term highway traffic flow. The experiments show that ARIMA (p, d, q) of order (0, 1, 3) can better model the given datasets. Many ARIMA-based models have been proposed for traffic prediction in the past decades. For example, ARIMA model of order (0, 1, 1) is proposed to reproduce all original time series. The experiments show that it is the most adequate model to reproduce all original time series [8].
If a seasonal component is added into ARIMA, we gain the structure denominated model called SARIMA. In [9], Williams et al. asserted that a one-week lagged first seasonal difference applied to discrete interval traffic data will yield a weakly stationary transformation. Resting on this assertion and the Wold decomposition theorem, Williams et al. presented a theoretical hypothesis that a univariate traffic data stream can be modeled as SARIMA. To validate the theoretical hypothesis, they performed experiments based on actual dataset from ITS and the empirical results are consistent with the theoretical hypothesis [9].
For traffic prediction, LR is one of the most typical non-parametric methods. Rice et al. discovered two naive predictors of travel time T T e ( t + δ ) : historical mean travel time μ T T ( t + δ ) and current status travel time T e ( t ) [10]. They concluded that travel time has a linear relation with its two naive predictors. Based on a dataset gathered from 116 single loop detectors along 48 miles in Los Angeles, they compared their method with other methods including principal components method and nearest neighbor method. The correctness of the conclusion is validated by the comparison.
In [11], k-NNR method is suggested as a candidate forecaster and is used for traffic prediction. The output value of k-NNR is the weighted average value of its k nearest neighbors on departure time. The empirical study describes the accuracy by comparing of the k-NN regression method to simple univariate time-series forecasts. In [12], k-NNR method is used to predict short-term traffic flow. In [13], k-NNR method is used to develop a model for dynamic multi-interval traffic volume prediction.
The idea of local weight is effective for traffic prediction. In [14], the effectiveness of local weight is validated. To suggest an approach for large-scale travel time prediction, Nikovski et al. presented an experimental comparison of several non-parametric methods [14], including LR, locally weighted regression, regression trees, k-NNR and neural networks. Although the non-linear methods have expected superiority over LR method, the locally weighted regression is the only non-linear method that can consistently outperform linear regression.
Given classification from the perspective of machine learning, the frequently used non-parametric methods usually fall into two categories: SVR and ANN.
The SVM method non-linearly maps input vectors to a very high dimension feature space, in which the linear decision surface for classification is constructed. SVM method does not depend on the dimension of input vectors. The high dimensionality space of SVM has high generalizability and big advantage for classification [15]. The SVR method is based on the SVM method. Thus, SVR-based methods are proposed for traffic prediction and achieve good performance. In [16], the SVR predictor with a radial basis function (RBF) kernel is proposed for travel time prediction. Based on the real highway traffic data, the proposed method achieves better performance than current-time predictor and historical-mean predictor. In [17], the incremental SVR method was proposed for traffic flow prediction. Based on the data sequentially collected by probe vehicles or loop detectors, the experimental results show that the proposed method is superior to the back-propagation neural network. In [18], an online version of SVR is proposed for short-term traffic flow prediction under atypical conditions (such as vehicular crashes, inclement weather, work zone and holidays).
Traffic prediction is based on modeling the complex non-linear spatiotemporal traffic dynamics in road network [1]. Many different types of neural networks have been proposed for traffic prediction including auto-encoders (AEs) [19], multi-layer perceptrons (MLP) [20] and recurrent neural networks [5,6,21].
In [19], Yisheng et al. rethought the traffic prediction problem based on deep architecture models, because existing traffic prediction methods are shallow and cannot live up to many real-world applications. Auto-Encoders (AEs) are proposed to learn generic traffic flow features and to predict traffic flow. AEs are trained on a greedy layer-wise way. Experimental results demonstrate that the proposed method for traffic flow prediction has superior performance compared to the baseline methods.
Polson et al. proposed a deep learning model for traffic flow prediction. The proposed model consists of a sequence of full-connection layers with activation function t a n h to extract features. The proposed model is the same as MLP. The experimental data are from 21 road segments. The road segments span thirteen miles and are the major corridor that connects Chicago’s southwest suburbs to its central business district [20]. Based on the experimental data, the effectiveness of the proposed model is validated. In experiments, the amount of input nodes is the amount of road segments.
The topology of neural network should be derived from traffic-related consideration. Thus, Elman RNN [22] is referred to as SSNN in [1]. The topology of SSNN consists of input layer, hidden layer, output layer and context layer. The input layer receives traffic data (such as traffic flow and average speeds) on the main carriage way, on-ramps and off-ramps (if available). The output layer consists of one neuron that calculates the predicted travel time. The context layer stores the previous internal states of the model. The hidden layer receives inputs from input layer and then stores them in context layer, finally transforming them into output layer.
CNN methods may improve the predictive accuracy by transforming traffic into images and using the implicit correlations in the nearest neighbors [2,3,4]. In [2], a CNN method is proposed for large-scale, network-wide traffic speed prediction. In [3], a fusion of CNN and LSTM is proposed for short-term passenger demand prediction. In [4], a CNN method with an Error-feedback RNN is proposed for continuous traffic speed prediction.
In [6], LSTM NN is proposed for traffic prediction. Based on the traffic speed data from two microwave traffic detectors that are deployed along the express-way without signal controls, the effectiveness of LSTM NN for traffic prediction is validated. In [5], LSTM NN is proposed for travel time prediction. Evaluation experiments are made based on the travel time dataset provided by Highways England. The experimental results show that travel time prediction error is relatively smaller than the baseline methods and 7.0% is the approximate median of the mean relative error of 66 links.
In summary, various techniques are involved in traffic prediction to improve the performance of traffic prediction. These prediction methods are proposed and evaluated based on specific traffic data separately, thus it is difficult to say which method is definitely superior over other methods in every situation. Neural networks can better capture complex non-linear spatiotemporal relationship. Neural networks, especially LSTM NN, are promising for traffic prediction.
The existing LSTM NNs for traffic prediction have two drawbacks: they do not use the departure time through the links for traffic prediction, and the way of modeling long-term dependence in time series is not direct in terms of traffic prediction. Thus, we assume that the way of modeling long-term dependence in time series may be improved by using more direct access. Travel time is correlated with departure time because traffic dynamics have some periodicity with departure time when traffic flow is not free. Thus, LSTM NN for traffic prediction may be improved by efficiently using of departure time.

2.2. Attention Mechanism

Attention mechanism has recently succeeded in image classification [23], neural machine translation [24], multimedia recommendation [25] and many other tasks, because it can concentrate on the effective parts of features adaptively.
In the tasks for image classification, it is computationally expensive to apply convolutional neural networks (CNN) on large images because the computational cost of CNN scales linearly with the number of pixels of input images. To address the problem of enormous computation cost, a novel RNN model with attention is presented in [23]. The attention mechanism helps the proposed model to adaptively select a sequence of regions from image or video and only processes the selected regions at high resolution.
In the tasks for machine translation, attentional mechanism selectively focuses on the effective parts of input sentences during translating and improves the accuracy of machine translation. Luong et al. proposed two approaches of attention mechanism for machine translation: a global approach that always attends to all source words and a local approach that only considers a subset of the source words at a time [24].
In the task of multimedia recommendation, existing collaborative filtering systems (CF) ignore the implicitness in the users’ interactions with multimedia content. In [25], a two-layer attention mechanism is proposed to extract implicit feedback. The bottom layer adaptively selects the informative implicit feedbacks on component level. The upper layer adaptively selects the informative implicit feedbacks on item level. The selected implicit feedbacks are incorporated into the classic CF model with implicit feedback.
In this paper, the attention mechanism is proposed to address the two drawbacks of LSTM NN. The traditional recurrent way to construct the depth of LSTM NN is substituted by attention mechanism. The attention mechanism is over the output layers of LSTM NN to model long-term dependence. The departure time is used as the aspect of the attention mechanism and the attention mechanism is used to integrate departure time into the proposed model.

3. Methodology

3.1. LSTM NN for Travel Time Prediction

The transition function of standard RNN, h t = t a n h ( W · [ h t 1 ; x t ] + b ) , is a linear layer followed by a point-wise non-linear function (such as hyperbolic tangent function). Standard RNN suffers from a problem of vanishing or exploding gradients. The gradients may grow or decay exponentially over long sequences [26,27]. LSTM NN is a type of RNN with more sophisticated and powerful transition ability to process the gradients; it is proposed to address the problem. In the remainder of this section, we first review the architecture of LSTM unit illustrated in Figure 1. Then, we review the LSTM NN illustrated in Figure 2 and its usage for travel time prediction. LSTM unit has three gates and a memory cell: input gate i t , forget gate f t , output gate o t , and memory cell c t . These gates adaptively keep or override information in the memory cell, forget previous information and decide how to access memory cell [27].
The transition functions of LSTM unit are described in Equations (1)–(6). W i , W f , W o R d × 2 d are the weight matrices of input gate, forget gate and output gate, respectively. b i , b f , b o R d are the biases of input gate, forget gate and output gate, respectively. W r is the weight matrix of memory cell and b r is the bias of the memory cell. These parameters are learned during LSTM NN training. σ is a sigmoid function. ⊙ denotes element-wise multiplication. x t is the input vector that is fed into the input layer of the LSTM unit on step t. h t 1 is the output vector of the hidden layer of the LSTM unit on step t 1 and is fed into the LSTM unit on step t. h t is the output vector of the hidden layer of the LSTM unit on step t. c t 1 denotes the memory cell of the LSTM unit on step t 1 . c t denotes the memory cell of the LSTM unit on step t.
i t = σ ( W i · [ h t 1 ; x t ] + b i )
f t = σ ( W f · [ h t 1 ; x t ] + b f )
o t = σ ( W o · [ h t 1 ; x t ] + b o )
g t = t a n h ( W r · [ h t 1 ; w t ] + b r )
c t = i t g t + f t c t 1
h t = o t t a n h ( c t )
In LSTM NN, LSTM units are placed end to end. LSTM units are used recursively to extract features from the complex non-linear spatiotemporal traffic data on road networks. LSTM NN recursively transits x t , h t 1 and c t 1 into the output vector h t . Finally, a vector h n is obtained that represents the predicted travel time.
An architecture of LSTM NN for travel time prediction is illustrated in Figure 2. { x 1 , x 2 , , x n } is an input sequence of LSTM NN. For travel time prediction, every item in the input sequence is a travel time value. LSTM NN uses the transition functions to extract { h 1 , h 2 , , h n } from the input sequence. { h 1 , h 2 , , h n } is the output vectors of the hidden layer of LSTM NN. h n represents the output value of LSTM NN. A predictor is usually linked to the end of LSTM NN h n . A single layer neural network is usually used to implement the predictor. Based on the predictor, vector h n is transited into travel time y ^ that is a real number. The predictor is formulated by Equation (7).
y ^ = W s · h n + b s

3.2. Attention Mechanism

Attention mechanism has recently demonstrated success in a wide range of tasks [23,24,25]. It is implemented by constructing a neural network according to the corresponding tasks. The neural network that is used for attention mechanism is usually called attentive neural network. An attentive neural network is illustrated in Figure 3. The transition functions of the attentive neural network are described by equations from Equations (8)–(10). H is an matrix and denotes the extracted features by prediction model, such as the aforementioned matrix [ h 1 , h 2 , , h n ] . e n R n is a vector of 1 s and v a is the embedding of aspect for attention mechanism. α is a vector and denotes the attention weights to features H. r is the final output of the attentive neural network and denotes the weighted sum of features H.
M = t a n h ( W h H W v v a e n
α = s o f t m a x ( w T M )
r = H α T
Based on aspect, attention mechanism effectively extracts the features in input sequences. Specifically, attention mechanism allows the model to attend over features based on an aspect, thereby it mitigates the model bottleneck. The attention mechanism lets the model capture the whole traffic dynamics in input sequence. The attention mechanism only informs our proposed model to accumulate the parts of features needed to attend over while reading the input sequence and accumulating to a representation in the cell state.

3.3. LSTM-Based Method with Attention Mechanism

Based on the aforementioned discussion, we propose an LSTM-based model with attention mechanism for travel time prediction in this section. The proposed model better captures the relationship between history travel time and predicted travel time. First, the proposed model is a tree structure and is not a chain structure. The LSTM units are not placed end to end in the proposed model. The attention mechanism is used to construct the depth of LSTM NN and the structural pattern is different from the unfold way of standard LSTM NN. The attention mechanism is over the output layer of each LSTM unit to model long-term dependence. The proposed method models long-term dependence in time series by direct and efficient travel time prediction. Second, the proposed model uses the departure time through the links for traffic prediction. The proposed model integrates departure time for applying some periodicity with departure time. The departure time is used as the aspect of the attention mechanism. If departure time aspect is employed, the attention mechanism allows the proposed model to capture more important part of input sequences than LSTM NN without departure time.
The architecture of the LSTM-based model with attention mechanism for travel time prediction is illustrated in Figure 4. In Figure 4, [ x 1 , x 2 , , x n ] denotes the input sequence into the proposed model. Let the departure time for the prediction be at 9 : 30 and n is 7, then the input sequence [ x 1 , x 2 , , x n ] is at [7:45, 8:00, 8:15, 8:30, 8:45, 9:00, 9:15].
Based on the given dataset the time interval (time period) is 15 min. Thus, one day contains 96 time intervals and 96 departure times. In the proposed model, the randomly initialized 96 embeddings for the departure times are from 0 to 96. The embeddings are learned during model training. Let departure time be at 9:30, then the index of the departure time in the embeddings is 38 and v 3 8 is selected as the departure time aspect of the attention mechanism. v 3 8 is v a in Equation (8). [ h 1 , h 2 , , h n ] in Figure 4 denotes H in Equation (8). [ a 1 , a 2 , , a n ] in Figure 4 is a in Equation (9) and are attention weights on features H. r in Figure 4 is computed using Equation (10). The end of the proposed model is a predictor that is a single layer neural network. The predictor for travel time value refers to Equation (12).
h s = t a n h ( W p r + W x h n )
y ^ = W s h s + b s

3.4. Models Training

We use a back-propagation method to train the proposed model. The back-propagation method is an end-to-end way method. The training is an optimization process to minimize the reward function (loss function) of the proposed model. In this paper, the reward function is an absolute loss function. We minimize the reward function value while the input sequences are sequentially fed into the proposed mode. The parameter set of the proposed model includes hyper-parameters, weight parameters and bias parameters. We choose the hyper-parameters of the proposed model on validation set, adjust the weights and biases of the proposed model on training set and evaluate the performance of the proposed model on test set. y is the observed travel time, y ^ is the corresponding predicted travel time, J is the cost value, and n is the number of the travel time sequences.
l o s s = a r g m i n Θ J ( y , y ^ )
J ( y , y p r e d ) = n ( y y ^ ) 2
In our experiments, we uses a mini-batch gradient descent optimization method, AdaGrad [28], to training the proposed model. AdaGrad adapts its learning rate to the parameters of the proposed model. AdaGrad performs much larger updates for infrequent parameters than frequent parameters and is well suit to deal with sparse data. Dean et al. [29] found that AdaGrad greatly improves the robustness of SGD. The rules of AdaGrad to update parameters are described by Equations (15)–(17).
θ t + 1 , i = θ t , i η G t , i i + ε · g t , i
g t , i = θ J ( θ i )
G t , i i = t g t , i 2
AdaGrad modifies the general learning rate of parameter θ i at time step t + 1 based on η and gradients θ i and G t at time step t. Here, G t R d × d is a diagonal matrix. Each element on the diagonal line of matrix G t , i i is a squared sum over g t , i . ε is a smoothing term that avoids division by zero (usually 1 × 10 8 ). Let learning rate η be 0.1. A main benefit of AdaGrad is that it eliminates the need to manually adjust the learning rate of model training [28].
Based on Figure 4 and its equations, the parameter set of the proposed model is { W i , b i , W f , b f , W o , b o , W c , b c , v , W h , W v , w , W p , W x , W s , b s } . To accelerate training, a 4-G GPU computer with CUDA [30] was used for the proposed model. The initial size of CNMeM was enabled with 80.0% memory. Theano was used for implement our proposed model.

4. Experiment

We adopted three evaluation indices to evaluate the travel time prediction accuracy of the proposed model: the mean absolute error (MAE), the mean absolute percentage error (MAPE), and the RMS error (RMSE). The evaluation indices are described by Equations (18)–(20). n is the length of the given test set, y is the observed travel time and y ^ is the predicted travel time.
M A E = 1 n i = 1 n | y i y ^ |
M A P E = 1 n i = 1 n | y i y ^ | y i
R M S E = [ 1 n i = 1 n ( y i y ^ ) 2 ] 1 2

4.1. Dataset

The experiments were performed on a dataset provided and managed by the Highways Agency in England [31]. This dataset provides average journey time (travel time) within 15-min time periods (time intervals) on all motorways and “A” roads. We selected a database from 1 March 2015 to 31 March 2015. The precision of travel time is 1 s. The dataset is described in Table 1, where one row is one time period.
The latest journey time and traffic data are available in [31]. This data series provides average journey time, speed and traffic flow information for 15-min periods since April 2009 on all morotways and “A” roads managed by the Highways Agency, known as the Strategic Road Network, in England. Journey times and speeds are estimated using a combination of sources, including Automatic Number Plate Recognition (ANPR) cameras, in-vehicle Global Positioning Systems (GPS) and inductive loops built into the road surface. Journey times are derived from real vehicle observations and imputed using adjacent time periods or the same time period on different days.

4.2. Task Definition

We predicted the travel time through links at time interval. Given one link selected, we predicted the travel time through the link at certain time interval. We divided the dataset that is related to the link into three parts: the data from 1 March to 21 March for the training set (approximately 87.1%), those on 28 March for the validation set (approximately 3.2%), and those from 29 March to 31 March for the test set (approximately 9.7%).
Input sequences and its corresponding outputs: The input sequences and its corresponding outputs were constructed from the database based on the selected link. The input sequences were fed into the proposed model and the corresponding outputs were observed values. The transition functions from the database to the input sequences are described in Equations (21) and (22).
L = r 1 r 2 r 3 r m f X = r 1 r 2 r w r 2 r 3 r w + 1 r m 1 y 1 y 2 y m w
y i = r w + i
Given link L, and r is the row in Table 1. Based on the task, The field average JT in Table 1 was selected as the item of input sequences. The values of average JTs were formatted into real numbers. Given the unfolded size of the proposed model is w, there are w LSTM units in the proposed model.
Matrix X represents the input sequences and its corresponding output for the proposed model and is generated from L by using Equations (21) and (22). Matrix X is fed into the the proposed model row by row for training or testing. For example, [ r 1 , r 2 , , r w ] is regarded as an input sequence, and y 1 = r w + 1 is regarded as the corresponding output.
There are 31 days in the selected database and 96 time periods per day (one time period is 15 min). Based on the selected database, there are m = 2976 ( 31 × 96 ) rows in matrix L and m w rows in matrix X.
Normalizing: Normalization: Normalization is a very effective method to prepare input sequences. Our experiments showed that normalization of the dataset is important for the training process and can avoid overflows in calculation. In our experiments, the input sequences and its corresponding output were rescaled to the range of 0 to 1 because the activation functions in the gates of LSTM units are sigmoid functions.

4.3. Parameters Setting for the Proposed Method

We needed to determine the amount of LSTM units in the proposed model because of its recursion. The amount also decides the amount of item in one input sequence that is the w in Equation (21). In the proposed model, let the amount of LSTM units be w and the input sequence R 1 × w . Let the dimension of h and o of LSTM unit be 4. Hyper-parameters for the proposed model are described in Table 2.
During validation process, we tried to obtain the optimal w of the proposed model. The definition domain of w ranges from 1 to 12. The experimental result is described in Figure 5. The optimal w was 7. The experimental result indicates that the amount of LSTM units should be sufficiently large to obtain a better accuracy. We obtained the hyper-parameters during validating process. The mini-batch size was 50. The initial learning rate of W s l r 1 was 0.1, and the initial learning rate of other parameters l r 2 was 0.01.
{ W i , b i , W f , b f , W o , b o , W c , b c , v , W h , W v , w , W p , W x , W s , b s } is the parameter set of the proposed model that is trained based on the given dataset. { W i , W f , W o , W c } R 4 × 5 are the weight matrices of the input gate, forget gate, output gate and memory cell in the proposed model, respectively, which are described in Equations (1)–(6). { b i , b f , b o , b c } R 4 are the respective biases described in Equations (1)–(6). { W h , W v } R 4 × 4 , w R 8 × 1 and v R 96 × 1 are the parameters for attention mechanism described in Equations (8) and (9). { W p , W x } R 4 × 4 are the parameters in Equation (11). W s R 4 and b s R 1 are the parameters for the final predictor described in Equation (12).

4.4. Parameters Setting for the Baseline Methods

The baseline methods in this study were: RW, LR, k-NNR, SARIMA, SVR, SSNN, CNN and LSTM NN.
The simplest baseline method is RW method. In the RW method, the traffic conditions at time interval t + δ are equal to traffic conditions at time interval t. Thus, the function of the RW baseline method is: X t + δ = X t , where X denotes travel time and t denotes time interval.
The LR method in our experiments is based on the work of Rice and Zwet [10]. In the literature, the learn rate of the parameters is set to 0.05. In our experiments, lag δ was set to 15 min and AdaGrad was used for the LR method. The parameter set of the LR method was {W1, W2, b1}. The function of the LR method was X t + δ = W 1 × X t + W 2 × μ T T ( t + δ ) + b 1 , where μ T T ( t + δ ) denotes historical mean travel time and is the historical mean of X t + δ in the all previous days.
A k-NNR method was used for travel time prediction in our experiments. The output value of the k-NNR is a weighted average value of its k nearest neighbors. The time intervals of the k nearest neighbors are before the time intervals of the predicted value. In the k-NNR for travel time prediction, the ways of giving different weights to k nearest neighbors are usually uniform and distance. The uniform method is that the weight values are equal. The distance method is that the weight values are the inverse of their distance to the predicted value. We let the neighbors of k-NNR be 7 and all neighbors were previous to the predicted travel time on time intervals.
We built time series { X i } by using Equations (21) and (22). We presumed that { X i } is an SARIMA (p,d,q) (P,D,Q) S process with period S. The SARIMA model was trained based on the literature [9]. The parameters of the SARIMA model in our experiments were set as follows: (p,d,q) was (1,0,1), (P,D,Q) was (0,1,1), and S was 96.
Linear kernel (linear), polynomial kernel (poly) and radical basis function kernel (rbf) were used as kernels of the SVR for travel time prediction in our experiments. The rbf kernel is proposed for traffic prediction in [16]. The independent term in poly and sigmoid kernel functions were set to 1 e 3 . The penalty parameter of the error term was set to 1.0. The degree of the polynomial kernel function was set to 3. 1 | f e a t u r e s | was used as kernel coefficient for SVR. The tolerance for stopping criterion was set to 1 e 3 .
The SSNN in our experiments was trained based on the work in [1]. The parameters of the SSNN in our experiment are described as follows. The topology for the SSNN in our experiments was that the size of input layers, hidden layers and output layers are all 7. The recurrent connections were fixed at 1.0 and not subject to adjustment. The context units were initially set to 0.5. The function was x c ( k ) = x ( k 1 ) . The iteration time was 25. The size of the hidden layer was 4. The batch size was 25. The parameter set was { W f , b f , W s , b s } , where W R 4 × 5 and b R 4 . The W f , b f were for RNN and the W s , b s sdf for the final predictor that is the same as Equation (7).
The CNN in our experiments was { W , b , W 1 , W 2 , b 2 , W s , b s } , where W , b is the convolution layer. The window of the convolution was 1 × 2 . W 1 / i n R 30 × 10 and W s / i n R 10 × 1 was for the connection layers. The parameters of the final predictor was W s , b s , which is the same as Equation (7). The iteration time was 40. The mini-batch was 25.
The LSTM NN was based on the work in [6]. The unfolded size of the LSTM NN was the same as the proposed model, i.e., 7. The iteration time was 25. The size of the hidden layer was 4. The batch size was 25. The parameter set was { W i , W o , W f , W c , b i , b o , b f , b c , W s , b s } , as described in Equationss (1)–(7), where W R 4 × 5 and b R 4 .

4.5. Similarity between the Prediction Value and the Observation Value

The observed travel times and the corresponding predicted travel times through three short length links are illustrated to compare the similarity of change in Figure 6. Three links are under low (AL1439), medium (AL2202) and heavy (AL1212) traffic loads respectively. As shown in Figure 6, the predicted travel times exhibit similar change patterns of travel time with the observed travel times. The changes are in good agreement in low, medium and heavy traffic conditions.
An medium length link L is illustrated in Figure 9. The observed travel times through link L are illustrated in Figure 7. The corresponding predicted travel times are also included in Figure 7 to compare the similarity of change. As shown in Figure 7, the predicted travel times exhibit similar change patterns of travel time with the observed travel times.

4.6. Accuracy Comparison Based on a Short Length Link

Based on link AL2202, the comparison between the proposed model and several baseline methods are described in this section. The length of AL2202 is 2.1 km.
Prediction accuracy: To evaluate the prediction accuracy of the proposed model, we compared the proposed model and the baseline methods based on short length link AL2202. The mean experimental results of these methods are given in Table 3. The experiment results show that the proposed model achieved better accuracy than the baseline methods. The proposed model outperformed the LSTM NN in terms of the accuracy and the proposed model also outperformed other baseline methods.
Comparison of convergence speed: Based on the link AL1896, Figure 8 illustrates a comparison of the convergence speed between the proposed model and the LSTM NN. The result shows that the proposed model had faster convergence speed than the LSTM NN. Thus, the attention mechanism was sufficient to extract features from the input sequences and accumulate to a representation in memory cell. The attention mechanism informed the LSTM units to addend over the efficient parts of features.

4.7. Accuracy Comparison Based on a Medium Length Link

We extended the comparison experiments based on an medium length link L on a road network. The tail link AL2291 and the head link AL3069A were removed from the road network to build a medium length Link L . Link L contained the links from AL3070 to AL1877, as described in Figure 9. The length of link L is approximately 54.38 km.
Prediction Accuracy: Based on link L , we compared the proposed model and the baseline methods for travel time prediction. The baseline methods and the corresponding setting are descried in the previous section. The comparable results is described in Table 4. The proposed model outperformed the LSTM NN and other baseline methods in terms of prediction accuracy. This conclusion based on the medium link L is consistent with the conclusion based on the shot length link AL2202.
Time cost: Based on medium length link L , we performed experiments to compare the time cost between the proposed model and the baseline methods. The experiments results are illustrated in Figure 10. The SARIMA method took the longest time to execute the training and testing. The methods based on NN (SSNN, LSTM NN and the proposed model) were likely to take more time training and testing than other none-NN methods (RW, k-NNR, SVR and LR) due to the complexity of the architecture and training of the NN-based methods. The time cost of the NN-based methods were within a reasonable boundary because of the rapid increase in the capacity of computers. The increase of the capability of computer promotes the recovery of NN-based methods, especially in the artificial intelligence domain.

4.8. A Further Evaluation on Link AL1167

The proposed model can be used for traffic prediction because there exist temporal relationship between the predicted travel time and the previous predicted travel time. To validate this assumption, we demonstrated the relationship by employing the Pearson correlation given by Equation (17), where X and Y denote two random variables with the same number of observations. The Pearson correlation was also used to explore the spatiotemporal correlations among the variables in short-term passenger demand prediction [3].
C o r r ( X , Y ) = E ( X Y ) E ( X ) E ( Y ) E ( X 2 ) E 2 ( X ) ( Y 2 ) E 2 ( Y )
We calculated the Pearson correlations between the predicted travel time and the previous predicted travel time. The predicted travel times was { T t } and the previous predicted travel times were { T t 1 } , { T t 2 } , { T t 3 } , , { T t 11 } , the time intervals of which are 1, 2, 3, ..., 11, respectively.
Figure 4 shows the correlations between { T t } and { T t 1 } , { T t 2 } , { T t 3 } , , { T t 11 } . It can be observed that the correlations between dropped gradually with the increase of time intervals. This means that there exited strong temporal correlations between the predicted travel times and its previous travel times. The correlations analysis validated the importance of the input sequences spanning time intervals and constructing the depth of LSTM to modeling long-term dependence.
To evaluate the effectiveness of the proposed model during periods of high volatility, we show the MAE (red dots) against the observed values (blue line) for three days on link AL1167 in Figure 11. We can see that the proposed model could capture the traffic regime changes, especially during high fluctuations, but the predictive power of the proposed model Was not uniform during the three days. There were four large errors in the 297 traffic predictions in Figure 11 when the traffic regime abruptly changed. The four large errors are marked in red. Nos. 1, 3 and 4 from left to right were observed at when traffic regime changed to congestion. No. 2 was observed when traffic regime changed to recovery.
To evaluate the predictive power of the proposed method to capture the traffic condition propagation in terms of travel time, we describe the travel time prediction results by using the heat maps on link AL1996 in Figure 12. The observed values are shown in the first plot and the predicted values are shown in the second plot. In contrast, we see that the proposed method properly captured both forward and backward congestion propagation during the three days.
Table 5 demonstrates the prediction performance of different methods with various time lags for AL1167. The method with the best performance is marked in bold. Based on AL1167, a general trend can be found that the prediction accuracies for the methods increased as the time lag became longer. We can see that the proposed model achieved the best prediction accuracy when time lag was 1. This conclusion is the same as the correlations in Figure 13.

4.9. Result Analysis

It is well known that the prediction errors of models can be decomposed into two main subcomponents: error due to “bias” and “variance”. The bias errors are the difference between prediction values and observed values. The bias errors are due to erroneous assumptions in the prediction model. High bias implicates under fitting that the models missed some relevant relations between features and predictive outputs. The variance error is a difference of bias error. The variance error is due to the sensitivity of model in the small fluctuations in training set. High variance implicates over fitting that the model transits the random noise in training data, rather than the intended outputs. There is usually a trade off between bias and variance [32].
Bias discussion: The proposed method outperformed all baseline methods in terms of bias. Thus, there was no under-fitting problem in the proposed model.
Traffic dynamic has some periodicity with departure time when traffic flow is not free. Thus, departure time was an effective input for traffic prediction. If the same input sequences were fed into LSTM NN, LSTM NN without departure time outputed the same predicted value even if the predicted values and their departure times were different. Obviously, LSTM NN without departure time was less effective for travel time prediction than the proposed model. The experiment results prove the hypothesis. Table 3 and Table 4 show that the proposed model has the smallest bias. The proposed model exhibited the best accuracy and can effectively predict travel time through the link. Specifically, using the departure time through the links in the proposed model was feasible and effective for improving LSTM NN for traffic time prediction. In the proposed model, departure time is used as the aspect of the attention mechanism and the attention mechanism is used to integrate departure time into the proposed model.
In the proposed model, the attention mechanism substitutes the traditional recurrent way to construct the depth of LSTM and model long-term dependence in time series. The new way is direct in terms of traffic prediction. The proposed model obtained the best performance and outperformed all other methods in terms of prediction accuracy. Thus, the new way is efficient to predict travel time through links.
Because link L is longer than link A L 2202 , the bias of the proposed model in Table 4 is smaller than it is in Table 3. This conclusion is the same as the conclusion in [16].
Variance Discussion:Table 4 shows that the variances of the proposed model is on the same order of magnitude as the baseline methods except the k-NN with uniform function and the SARIMA method. The proposed method outperformed seven of the baseline methods in terms of variance. Thus, there was no over-fitting problem in the proposed model.

4.10. Case Study

To validate the effectiveness of applying departure time and the new way to construct the depth of the proposed method, the experiments were repeated twice based on link AL2202. Let 10:00 be the departure time through link AL2202. The departure time of 10:00 was chosen at random. The predicted travel times and their attention weights are recorded in Table 6 and Table 7.
Method stability: We compared the prediction accuracy MAPE of the proposed method between Table 6 and Table 7. The results on 29 March indicate that the prediction accuracy was not affected by the training process. Thus, the proposed method was stable. We came to the same conclusion on 30 and 31 March.
The prediction of travel time change tendency: We compared the travel time at 10:00 with those at 9:45 on 29 March (Table 6); the observed travel time at 10:00 and the predicted travel time at 10:00 were shorter than the observed travel time at 9:45. In conclusion, the predicted travel time at 10:00 and the observed travel time at 10:00 had the same decreasing trend in their change.
We compared the travel time at 10:00 and those at 9:45 on 30 March (Table 6); the observed travel time at 10:00 and the predicted time at 10:00 were longer that the observed travel time at 9:45. In conclusion, the predicted travel time at 10:00 and the observed travel time at 10:00 had the same increasing trend in their change.
We compared the travel time at 10:00 and those at 9:45 on 31 March (Table 6); the predicted travel time at 10:00 and its observed travel time did not show the same change trend. The observed travel time at 10:00 was longer than the travel time at 9:45 and the predicted travel time at 10:00 was shorter than its travel time at 9:45. Perhaps, the proposed model could regress toward the mean, because the predicted travel time was too long.
The change trend of the predicted travel time or the regression toward the mean presented in Table 7 was the same as those in Table 6.
Efficiency of the departure time aspect:Table 6 and Table 7 show that the attention mechanism focused on the output features of the proposed model given a departure time. In the tables, the attention weights indicate the importance degree of the output features: the larger is the weight, the more important is the feature. If the predicted travel time were larger than its corresponding one-step travel time, the attention weights would be more focused on the the larger travel time. If the predicted travel time were smaller than the corresponding one-step travel time, the attention weights would be more focused on the small travel time. In conclusion, the attention mechanism could effectively focus on the output features to produce the weighted sum r in Figure 4 when dynamically given a departure time. It is an effective new way to construct the depth of the proposed method.

5. Conclusions

In this paper, we propose an LSTM-based method with attention mechanism for travel time prediction. We use the attention mechanism to construct the depth of the proposed model. The departure time is used as the aspect of the attention mechanism and is integrated into the proposed model. Comparison experiments were performed based on the database provided by Highways England. The experimental results show that the proposed model obtained better accuracy than the baseline model. The case study suggests that the attention mechanism can effectively concentrate on the differences of input features. Thus, the proposed model is feasible and effective.
As future work, we will improve the trend and avoid regression toward the mean. A potential direction would be the application of a local optimization approach.

Author Contributions

Conceptualization, X.R., Z.S. and Y.F.; Funding acquisition, Z.S.; Investigation, X.R.; Methodology, X.R., Z.S. and Y.F.; Project administration, Z.S.; Supervision, Z.S. and C.L.; Validation, X.R.; Writing—original draft, X.R.; and Writing—review and editing, Y.F.

Funding

This research was funded in part by the National Key R & D Program of China (grant numbers 2018YFB0803400 and 2018YFB0505302) and the National Science Foundation of China (grant number 61832012).

Acknowledgments

We are thankful to Jicheng Zhou for the help provided with laboratory equipment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lint, H. Reliable Travel Time Prediction for Freeways. Ph.D. Thesis, The Netherlands TRAIL Research School, Delft, The Netherlands, May 2004. [Google Scholar]
  2. Ma, X.; Dai, Z.; He, Z.; Ma, J.; Wang, Y.; Wang, Y. Learning traffic as images: A deep convolutional neural network for large-scale transportation network speed prediction. Sensors 2017, 17, 818. [Google Scholar] [CrossRef] [PubMed]
  3. JKe, J.; Zheng, H.; Yang, H.; Chen, X.M. Short-term forecasting of passenger demand under on-demand ride services: A spatio-temporal deep learning approach. Transp. Res. Part C Emerg. Technol. 2017, 85, 591–608. [Google Scholar] [Green Version]
  4. Wang, J.; Gu, Q.; Wu, J.; Liu, G.; Xiong, Z. Traffic speed prediction and congestion source exploration: A deep learning method. In Proceedings of the IEEE International Conference on Data Mining, Barcelona, Spain, 12–15 December 2016; pp. 499–508. [Google Scholar]
  5. Duan, Y.; Lv, Y.; Wang, F.Y. Travel time prediction with LSTM neural network. In Proceedings of the IEEE International Conference on Intelligent Transportation Systems, Rio de Janeiro, Brazil, 1–4 November 2016; pp. 1053–1058. [Google Scholar]
  6. Ma, X.; Tao, Z.; Wang, Y.; Yu, H.; Wang, Y. Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp. Res. Part C 2015, 54, 187–197. [Google Scholar] [CrossRef]
  7. Ahmed, M.S.; Cook, A.R. Analysis of Freeway Traffic Time-Series Data by Using Box-Jenkins Techniques; Transportation Research Board: Washington, DC, USA, 1979; pp. 1–9. [Google Scholar]
  8. Hamed, M.M.; Al-Masaeid, H.R.; Said, Z.M.B. Short-Term Prediction of Traffic Volume in Urban Arterials. J. Transp. Eng. 1995, 121, 249–254. [Google Scholar] [CrossRef]
  9. Williams, B.M.; Hoel, L.A. Modeling and Forecasting Vehicular Traffic Flow as a Seasonal ARIMA Process: Theoretical Basis and Empirical Results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef]
  10. Rice, J.; Van Zwet, E. A simple and effective method for predicting travel times on freeways. IEEE Trans. Intell. Transp. Syst. 2004, 5, 200–207. [Google Scholar] [CrossRef]
  11. Davis, G.A.; Nihan, N.L. Nonparametric Regression and ShortTerm Freeway Traffic Forecasting. J. Transp. Eng. 1991, 117, 178–188. [Google Scholar] [CrossRef]
  12. Smith, B.L.; Williams, B.M.; Oswald, R.K. Comparison of parametric and nonparametric models for traffic flow forecasting. Transp. Res. Part C Emerg. Technol. 2002, 10, 303–321. [Google Scholar] [CrossRef]
  13. Chang, H.; Lee, Y.; Yoon, B.; Baek, S. Dynamic near-term traffic flow prediction: Systemoriented approach based on past experiences. IET Intell. Transp. Syst. 2012, 6, 292–305. [Google Scholar] [CrossRef]
  14. Nikovski, D.; Nishiuma, N.; Goto, Y.; Kumazawa, H. Univariate short-term prediction of road travel times. In Proceedings of the Intelligent Transportation Systems, Vienna, Austria, 16 September 2005; pp. 1074–1079. [Google Scholar]
  15. Drucker, H.; Burges, C.J.C.; Kaufman, L.; Smola, A.J.; Vapnik, V. Support Vector Regression Machines. Adv. Neural Inf. Process. Syst. 1996, 28, 779–784. [Google Scholar]
  16. Wu, C.H.; Ho, J.M.; Lee, D.T. Travel-time prediction with support vector regression. IEEE Trans. Intell. Transp. Syst. 2004, 5, 276–281. [Google Scholar] [CrossRef]
  17. Su, H.; Zhang, L.; Yu, S. Short-term Traffic Flow Prediction Based on Incremental Support Vector Regression. In Proceedings of the International Conference on Natural Computation, Haikou, China, 24–27 August 2007; pp. 640–645. [Google Scholar]
  18. Castro-Neto, M.; Jeong, Y.S.; Jeong, M.K.; Han, L.D. Online-SVR for short-term traffic flow prediction under typical and atypical traffic conditions. Expert Syst. Appl. Int. J. 2009, 36, 6164–6173. [Google Scholar] [CrossRef]
  19. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic Flow Prediction With Big Data: A Deep Learning Approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar] [CrossRef]
  20. Polson, N.G.; Sokolov, V.O. Deep learning for short-term traffic flow prediction. Transp. Res. Part C Emerg. Technol. 2017, 79, 1–17. [Google Scholar] [CrossRef] [Green Version]
  21. Lint, J.W.C.V. Reliable Real-Time Framework for Short-Term Freeway Travel Time Prediction. J. Transp. Eng. 2006, 132, 921–932. [Google Scholar] [CrossRef]
  22. Elman, J. Finding structure in Time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  23. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst. 2014, 3, 2204–2212. [Google Scholar]
  24. Luong, M.T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. arXiv, 2015; arXiv:1508.04025. [Google Scholar]
  25. Chen, J.; Zhang, H.; He, X.; Liu, W.; Liu, W.; Chua, T.S. Attentive Collaborative Filtering: Multimedia Recommendation with Item- and Component-Level Attention. In Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 335–344. [Google Scholar]
  26. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 2002, 5, 157–166. [Google Scholar] [CrossRef]
  27. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  28. Duchi, J.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 257–269. [Google Scholar]
  29. Dean, J.; Corrado, G.S.; Monga, R.; Chen, K.; Devin, M.; Le, Q.V.; Mao, M.Z.; Ranzato, M.; Senior, A.; Tucker, P. Large scale distributed deep networks. In Proceedings of the International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1223–1231. [Google Scholar]
  30. Vanhoucke, V.; Mao, M.Z. Improving the speed of neural networks on CPUs. In Proceedings of the NIPS 2011 Workshop on Deep Learning and Unsupervised Feature Learning, Granada, Spain, 16–17 December 2011. [Google Scholar]
  31. Highways England. Highways Agency Network Journey Time and Traffic Flow Data. Available online: http://data.gov.uk/dataset/highways-england-network-journey-time-and-traffic-flow-data (accessed on 10 April 2017).
  32. Fortmannroe, S. Understanding the Bias-Variance Tradeoff. 2012. Available online: http://scott.fortmann-roe.com/docs/BiasVariance.html (accessed on 5 June 2017).
Figure 1. An architecture of LSTM unit.
Figure 1. An architecture of LSTM unit.
Sensors 19 00861 g001
Figure 2. An architecture of LSTM NN for travel time prediction.
Figure 2. An architecture of LSTM NN for travel time prediction.
Sensors 19 00861 g002
Figure 3. An architecture of attentive neural network.
Figure 3. An architecture of attentive neural network.
Sensors 19 00861 g003
Figure 4. Architecture of the LSTM-based model with attention mechanism for travel time prediction.
Figure 4. Architecture of the LSTM-based model with attention mechanism for travel time prediction.
Sensors 19 00861 g004
Figure 5. A change of MAPE when the amount of LSTM units in the proposed model increase. The experiment is made based on AL1896.
Figure 5. A change of MAPE when the amount of LSTM units in the proposed model increase. The experiment is made based on AL1896.
Sensors 19 00861 g005
Figure 6. Comparisons between the predicted travel times and its observed travel times through the links under different traffic conditions: (a) a comparison under low traffic conditions; (b) a comparison under medium traffic conditions; and (c) a comparison under heavy traffic conditions.
Figure 6. Comparisons between the predicted travel times and its observed travel times through the links under different traffic conditions: (a) a comparison under low traffic conditions; (b) a comparison under medium traffic conditions; and (c) a comparison under heavy traffic conditions.
Sensors 19 00861 g006
Figure 7. A comparison between the predicted travel times and the observed travel times through an medium length link L .
Figure 7. A comparison between the predicted travel times and the observed travel times through an medium length link L .
Sensors 19 00861 g007
Figure 8. A comparison of convergence speed between the proposed model and the LSTM NN. The comparison is based on Link AL1896. The unfold size of the LSTM NN and the proposed model were 7.
Figure 8. A comparison of convergence speed between the proposed model and the LSTM NN. The comparison is based on Link AL1896. The unfold size of the LSTM NN and the proposed model were 7.
Sensors 19 00861 g008
Figure 9. A road network consists of the links AL3069A, AL3070, AL2202, AL1900, AL1896, AL1891, AL1885, AL1883, AL1877 and AL2991 in Figure 9. The medium length link L consists of the links from link AL3070 to link AL 1877. A comparison between the proposed model and the proposed model was made based on link L .
Figure 9. A road network consists of the links AL3069A, AL3070, AL2202, AL1900, AL1896, AL1891, AL1885, AL1883, AL1877 and AL2991 in Figure 9. The medium length link L consists of the links from link AL3070 to link AL 1877. A comparison between the proposed model and the proposed model was made based on link L .
Sensors 19 00861 g009
Figure 10. A comparison of time cost between the proposed model and the baseline methods.
Figure 10. A comparison of time cost between the proposed model and the baseline methods.
Sensors 19 00861 g010
Figure 11. The residuals of the test set on link AL1996 over time. The horizontal axis is time interval and the vertical axis is travel time. The horizontal axis contains time intervals of three days and there are 96 time intervals per day.
Figure 11. The residuals of the test set on link AL1996 over time. The horizontal axis is time interval and the vertical axis is travel time. The horizontal axis contains time intervals of three days and there are 96 time intervals per day.
Sensors 19 00861 g011
Figure 12. Heat plots of travel times during three days. The horizontal axis are time intervals that spans three days and 96 time intervals per day. The color depth expresses the travel time value on link AL1996.
Figure 12. Heat plots of travel times during three days. The horizontal axis are time intervals that spans three days and 96 time intervals per day. The color depth expresses the travel time value on link AL1996.
Sensors 19 00861 g012
Figure 13. Travel time correlations on link AL1167 with length of 18.48 km. The best results in the table appear in bold.
Figure 13. Travel time correlations on link AL1167 with length of 18.48 km. The best results in the table appear in bold.
Sensors 19 00861 g013
Table 1. Data description of the database provided by the Highways Agency in England.
Table 1. Data description of the database provided by the Highways Agency in England.
Link RefDateTime Period (0–95)Average JT (s)Link Length (km)
AL18961 March 20140626.1218.32
AL18961 March 20141612.6018.32
AL18961 March 20142604.2318.32
Table 2. Hyper-parameters for the proposed model.
Table 2. Hyper-parameters for the proposed model.
VariableDescriptionValue
wThe unfold size of the proposed model7
dThe dimension of hidden layers and output layers of LSTM4
m i The mini-batch size50
l r 1 The initial learning rate of W s 0.1
l r 2 The initial learning rate of other parameters0.01
n o r m The range of the dataset normalization(0, 1)
Table 3. A comparison between the proposed model and the baseline methods based on short length link AL2202. The best results in the table appear in bold.
Table 3. A comparison between the proposed model and the baseline methods based on short length link AL2202. The best results in the table appear in bold.
Models Departure Time in 15 min
RMSEMAEMAPE (%)
RW 9.7872.56488.87
k-NNRuniform9.8002.5668.73
distance9.8002.5668.73
SARIMA 8.6372.5829.01
SVRlinear8.2022.4878.30
rbf8.1802.4868.30
poly8.2162.4978.36
LR 10.3232.7868.92
CNN 9.2702.7168.41
SSNN 9.2052.6508.37
LSTM NN 8.4722.4857.34
LSAM 8.1332.3567.01
Table 4. A accuracy comparison between the proposed model and the baseline methods based on medium length link L . The best results in the table appear in bold.
Table 4. A accuracy comparison between the proposed model and the baseline methods based on medium length link L . The best results in the table appear in bold.
Models 15-min on Training Set15-min on Test SetVariance (%)
RMSEMAEMAPE (%)RMSEMAEMAPE (%)
RW 87.4998.0183.3378.77707.54452.990.34
k-NNRuniform77.9557.6453.0378.8547.5452.990.04
distance14.3131.6630.1478.8547.5452.992.85
SARIMA 413.93119.96320.5193.2718.5083.8516.66
SVRlinear80.6287.7783.1472.9837.4082.900.24
rbf80.6467.7753.1473.0367.4272.920.22
poly84.0167.8923.2574.6707.5713.040.21
LR 83.9637.8972.9692.8808.5653.450.49
CNN 83.0937.8783.2072.1317.3992.870.33
SSNN 83.6317.8733.1872.0607.3612.840.34
LSTM NN 80.8857.7813.1471.6317.3182.830.31
LSAM 79.4987.6502.8170.1467.1882.690.22
Table 5. The results of the proposed method with various time lag based on AL1167.
Table 5. The results of the proposed method with various time lag based on AL1167.
ModelsTime Lag 1Time Lag 2Time Lag 3Time Lag 4
MAEMAPEMAEMAPEMAEMAPEMAEMAPE
CNN6.4326.876.9477.927.0808.287.2678.75
SSNN6.3386.106.9127.877.2658.707.4159.03
LSTM NN5.9795.956.8587.787.1978.477.3428.98
LSAM5.7885.616.7967.667.0698.287.2948.76
Table 6. Three experimental results collected in the first experiment of the proposed model.
Table 6. Three experimental results collected in the first experiment of the proposed model.
Departure Time8:158:308:459:009:159:309:4510:00
Steps Ahead7-Step6-Step5-Step4-Step3-Step2-Step1-StepObservedPredictedMAPE (%)
03-29AverageJT70.8271.8279.1078.7783.3083.3083.3074.3282.7711.4
Weights0.2270.2220.1420.1330.0960.0900.089
03-30AverageJT86.3886.9191.7177.7379.4789.0483.7396.8784.8012.5
Weights0.1480.1180.0830.1990.2030.1090.140
03-31AverageJT83.6385.3384.5789.3887.1091.9995.7897.2191.056.3
Weights0.2220.1700.1700.1240.1370.1000.076
Table 7. Three experiment results collected in the second experiment of the proposed model.
Table 7. Three experiment results collected in the second experiment of the proposed model.
Time Period8:158:308:459:009:159:309:4510:00
Steps Ahead7-Step6-Step5-Step4-Step3-Step2-Step1-StepObservedPredictedMAPE (%)
03-29AverageJT70.8271.8279.1078.7783.3083.3083.3074.3282.8011.4
Weights0.1810.1790.1400.1400.1200.1200.120
03-30AverageJT86.3886.9191.7177.7379.4789.0483.7396.8784.5412.7
Weights0.1340.1210.1040.1890.1800.1240.147
03-31AverageJT83.6385.3384.5789.3887.1091.9995.7897.2190.816.6
weights0.1700.1500.1610.1370.1500.1250.107

Share and Cite

MDPI and ACS Style

Ran, X.; Shan, Z.; Fang, Y.; Lin, C. An LSTM-Based Method with Attention Mechanism for Travel Time Prediction. Sensors 2019, 19, 861. https://doi.org/10.3390/s19040861

AMA Style

Ran X, Shan Z, Fang Y, Lin C. An LSTM-Based Method with Attention Mechanism for Travel Time Prediction. Sensors. 2019; 19(4):861. https://doi.org/10.3390/s19040861

Chicago/Turabian Style

Ran, Xiangdong, Zhiguang Shan, Yufei Fang, and Chuang Lin. 2019. "An LSTM-Based Method with Attention Mechanism for Travel Time Prediction" Sensors 19, no. 4: 861. https://doi.org/10.3390/s19040861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop