Abstract
Traffic flow prediction is a very challenging task in traffic networks. Travel order quantity prediction is of great value to the analysis of traffic flow. However, the number of travel orders is closely related to time, leading to its short-term proximity and long-term cyclical dependence. Therefore, it is difficult to predict travel order quantity using traditional methods. To capture both the long-term and short-term correlations, in this paper, we propose an attention-based bidirectional long short-term memory network (AT-BLSTM) model, which consists of bidirectional long short-term memory (BLSTM) layer and attention layer. The BLSTM layer contains forward and backward long short-term memory (LSTM), which uses a novel method to combine forward and backward output. The attention layer utilizes novel self-attention algorithm to assign different weights according to the correlation between features. In this case, AT-BLSTM can predict travel order quantity more accurately than other time series models. For example, AT-BLSTM achieves the lowest MAE (0.0647), the lowest RMSE (0.0836) and the lowest MAPE (0.1239) among all the methods on dataset Xiuying. Extensive experiments on real-world travel datasets offer evidence that the proposed approach matches or outperforms state-of-the-art methods.
Similar content being viewed by others
References
Malvin N, Pudjaprasetya SR (2021) Staggered conservative scheme for simulating the emergence of a Jamiton in a phantom traffic jam. Int J Intell Transp Syst Res 19(1):128–140. https://doi.org/10.1007/s13177-020-00229-y
Yin X, Wu G, Wei J, Shen Y, Qi H, Yin B (2021) Multi-stage attention spatial-temporal graph networks for traffic prediction. Neurocomputing 428:42–53. https://doi.org/10.1016/j.neucom.2020.11.038
Bai L, Yao L, Kanhere SS, Yang Z, Chu J, Wang X (2019a) Passenger demand forecasting with multi-task convolutional recurrent neural networks. Adv Knowl Discov Data Min 11440:29–42. https://doi.org/10.1007/978-3-030-16145-3_3
Bai L, Yao L, Kanhere SS, Wang X, Sheng QZ (2019b) Passenger demand forecasting with multi-task convolutional recurrent neural networks. Adv Knowl Discov Data Min 11440:29–42. https://doi.org/10.1007/978-3-030-16145-3_3
Zhang J, Zheng Y, Qi D, Li R, Yi X (2016) STG2Seq: spatial-temporal graph to sequence model for multi-step passenger demand forecasting. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Macao, China, pp 1981–1987. https://doi.org/10.24963/ijcai.2019/274
Zhang J, Zheng Y, Qi D (2017) DNN-based prediction model for spatio-temporal data. In: Proceedings of the 24th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems. ACM, Burlingame, California, pp 1–4. https://doi.org/10.1145/2996913.2997016
Zhang K, Feng Z, Chen S, Huang K, Wang G (2016) Deep spatio-temporal residual networks for citywide crowd flows prediction. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 1655–1661
Wang D, Cao W, Xu M, Li J (2016) ETCPS: an effective and scalable traffic condition prediction system. In: International Conference on Database Systems for Advanced Applications. Springer International Publishing, Cham, vol 9643, pp 419–436. https://doi.org/10.1007/978-3-319-32049-6_26
Li X, Pan G, Wu Z, Qi G, Li S, Zhang D, Zhang W, Wang Z (2012) Prediction of urban human mobility using large-scale taxi traces and its applications. Front Comput Sci 6(1):111–121. https://doi.org/10.1007/s11704-011-1192-6
Moreira-Matias L, Gama J, Ferreira M, Mendes-Moreira J, Damas L (2013) Predicting taxi-passenger demand using streaming data. IEEE Trans Intell Transp Syst 14(3):1393–1402. https://doi.org/10.1109/TITS.2013.2262376
Zhou X, Shen Y, Zhu Y, Huang L (2018) Predicting multi-step citywide passenger demands using attention-based neural networks. In: Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining. ACM, Marina Del Rey, pp 736–744. https://doi.org/10.1145/3159652.3159682
Huang W, Song G, Hong H, Xie K (2014) Deep architecture for traffic flow prediction: deep belief networks with multitask learning. IEEE Trans Intell Transp Syst 15(5):2191–2201
Zhao N, Ye Z, Pei Y, Liang YC, Niyato D (2020) Spatial-temporal attention-convolution network for citywide cellular traffic prediction. IEEE Commun Lett 24(11):5
Poole MA, OÃFarrell PN (1971) The assumptions of the linear regression model. Trans Inst Br Geogr 52:145–158
Ogutu JO, Schulz-Streeck T, Piepho HP (2012) Genomic selection using regularized linear regression models: ridge regression, lasso, elastic net and their extensions. BMC Proc 6(S2):S10. https://doi.org/10.1186/1753-6561-6-S2-S10
Granger C, Andersen A (1978) On the invertibility of time series models. Stoch Process Appl 8(1):87–92. https://doi.org/10.1016/0304-4149(78)90069-8
Asteriou D, Hall SG (2016) ARIMA models and the Box–Jenkins methodology. Appl Econ. https://doi.org/10.1057/978-1-137-41547-9_13
Li Y, Zheng Y, Zhang H, Chen L (2015) Traffic prediction in a bike-sharing system. In: Proceedings of the 23rd SIGSPATIAL International Conference on Advances in Geographic Information Systems. ACM, Seattle, pp 1–10. https://doi.org/10.1145/2820783.2820837
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3th International Conference on Learning Representations
Saon G, Tuske Z, Bolanos D, Kingsbury B (2021) Advancing RNN transducer technology for speech recognition. In: ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Toronto, pp 5654–5658. https://doi.org/10.1109/ICASSP39728.2021.9414716
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444. https://doi.org/10.1038/nature14539
Elman JL (1991) Distributed representations, simple recurrent networks, and grammatical structure. Mach Learn 7:195–225. https://doi.org/10.1007/bf00114844
Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1(2):270–280. https://doi.org/10.1162/neco.1989.1.2.270
Cho K, van Merrienboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, Doha, pp 1724–1734. https://doi.org/10.3115/v1/D14-1179
Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:141235551412.3555
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Ma X, Tao Z, Wang Y, Yu H, Wang Y (2015) Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp Res Part C Emerg Technol 54:187–197. https://doi.org/10.1016/j.trc.2015.03.014
Yu R, Li Y, Shahabi C, Demiryurek U, Liu Y (2017) Deep learning: a generic approach for extreme condition traffic forecasting. In: Proceedings of the 2017 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics, Philadelphia, pp 777–785. https://doi.org/10.1137/1.9781611974973
Yao H, Wu F, Ke J, Tang X, Jia Y, Lu S, Gong P, Ye J, Li Z (2018) Deep multi-view spatial-temporal network for taxi demand prediction. Proc AAAI Conf Artif Intell 32:2588–2595
Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5(2):157–166. https://doi.org/10.1109/72.279181
Gehring J, Auli M, Grangier D, Dauphin Y (2017) A convolutional encoder model for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Vancouver, pp 123–135. https://doi.org/10.18653/v1/P17-1012
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30:5998–6008
Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y (2018) Graph attention networks. In: 6th International Conference on Learning Representations
Cheng J, Dong L, Lapata M (2016) Long short-term memory-networks for machine reading. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/D16-1053
Zhou P, Shi W, Tian J, Qi Z, Li B, Hao H, Xu B (2016) Attention-based bidirectional long short-term memory networks for relation classification. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, Berlin, pp 207–212. https://doi.org/10.18653/v1/P16-2034
Lin Z, Feng M, dos Santos CN, Yu M, Xiang B, Zhou B, Bengio Y (2017) A structured self-attentive sentence embedding. In: 5th International Conference on Learning Representations
Kipf TN, Welling M (2017) Semi-supervised classification with graph convolutional networks. In: 5th International Conference on Learning Representations
Qin Y, Song D, Chen H, Cheng W, Jiang G, Cottrell G (2017) A dual-stage attention-based recurrent neural network for time series prediction. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, pp 2627–2633. https://doi.org/10.24963/ijcai.2017/366
Huang L, Ma Y, Wang S, Liu Y (2019) An attention-based spatiotemporal LSTM network for next POI recommendation. IEEE Trans Serv Comput. https://doi.org/10.1109/TSC.2019.2918310
Fernando T, Denman S, Sridharan S, Fookes C (2018) Soft + hardwired attention: an LSTM framework for human trajectory prediction and abnormal event detection. Neural Netw 108:466–478
Liu Q, Wang B, Zhu Y (2020) Short-term traffic speed forecasting based on attention convolutional neural network for arterials. Comput Aided Civ Infrastruct Eng 33(11):999–1016. https://doi.org/10.1111/mice.12417
Chung JH, Kim DW, Kang TK, Lim MT (2020) Traffic sign recognition in harsh environment using attention based convolutional pooling neural network. Neural Process Lett 51(3):2551–2573. https://doi.org/10.1007/s11063-020-10211-0
Yang T, Nan Z, Zhang H, Chen S, Zheng N (2020) Traffic agent trajectory prediction using social convolution and attention mechanism. In: 2020 IEEE Intelligent Vehicles Symposium (IV), pp 278–283. https://doi.org/10.1109/IV47402.2020.9304645
Guo G, Yuan W (2020) Short-term traffic speed forecasting based on graph attention temporal convolutional networks. Neurocomputing 410:387–393. https://doi.org/10.1016/j.neucom.2020.06.001
Xie Y, Xiong Y, Zhu Y (2020) SAST-GNN: a self-attention based spatio-temporal graph neural network for traffic prediction. In: Database Systems for Advanced Applications. Springer International Publishing, Cham, pp 707–714
Fang S, Zhang Q, Meng G, Xiang S, Pan C (2019) GSTNet: global spatial-temporal network for traffic flow prediction. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Macao, China, pp 2286–2293. https://doi.org/10.24963/ijcai.2019/317
Chang B, Park Y, Park D, Kim S, Kang J (2018) Content-aware hierarchical point-of-interest embedding model for successive POI recommendation. In: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Stockholm, Sweden, pp 3301–3307. https://doi.org/10.24963/ijcai.2018/458
Zhao P, Zhu H, Liu Y, Xu J, Li Z, Sheng VS, Zhou X (2020) Where to go next: a spatio-temporal gated network for next POI recommendation. IEEE Trans Knowl Data Eng
Wang D, Cao W, Li J, Ye J (2017) DeepSD: supply-demand prediction for online car-hailing services using deep neural networks. In: 2017 IEEE 33rd International Conference on Data Engineering (ICDE). IEEE, San Diego, pp 243–254. https://doi.org/10.1109/ICDE.2017.83
Lv Y, Duan Y, Kang W, Li Z, Wang FY (2014) Traffic flow prediction with big data: a deep learning approach. IEEE Trans Intell Transp Syst 16(2):865–873. https://doi.org/10.1109/TITS.2014.2345663
Duan Y, LV Y, Wang FY (2016) Travel time prediction with LSTM neural network. In: 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC). IEEE, Rio de Janeiro, pp 1053–1058. https://doi.org/10.1109/ITSC.2016.7795686
Ran X, Shan Z, Fang Y, Lin C (2019) An LSTM-based method with attention mechanism for travel time prediction. Sensors 19(4):861. https://doi.org/10.3390/s19040861
Petersen NC, Rodrigues F, Pereira FC (2019) Multi-output bus travel time prediction with convolutional LSTM neural network. Expert Syst Appl 120:426–435. https://doi.org/10.1016/j.eswa.2018.11.028
Zhang J, Wen Y, Long M, Wang J, Wang H (2020) Predictive recurrent networks for seasonal spatiotemporal data with applications to urban computing. Chin J Comput 43(02):286–302
Yan X, Mou L, Li G, Chen Y, Peng H, Jin Z (2015) Classifying relations via long short term memory networks along shortest dependency path. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp 1785–1794
Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: 3th International Conference on Learning Representations
Hoang ND, Nguyen QL (2019) A novel method for asphalt pavement crack classification based on image processing and machine learning. Eng Comput 35(2):487–498. https://doi.org/10.1007/s00366-018-0611-9
Acknowledgements
This work is supported by the National Natural Science Foundation of China (Grant No. 61772386), and the Open Fund of Key Laboratory of Urban Land Resources Monitoring and Simulation, Ministry of Natural Resources (Grant No. KF-2020-05-014).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service and/or company that could be construed as influencing that position in, or the review of, the manuscript entitled.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yang, F., Zhang, H. & Tao, S. Travel order quantity prediction via attention-based bidirectional LSTM networks. J Supercomput 78, 4398–4420 (2022). https://doi.org/10.1007/s11227-021-04032-8
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-021-04032-8