Skip to main content
Log in

Travel order quantity prediction via attention-based bidirectional LSTM networks

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Traffic flow prediction is a very challenging task in traffic networks. Travel order quantity prediction is of great value to the analysis of traffic flow. However, the number of travel orders is closely related to time, leading to its short-term proximity and long-term cyclical dependence. Therefore, it is difficult to predict travel order quantity using traditional methods. To capture both the long-term and short-term correlations, in this paper, we propose an attention-based bidirectional long short-term memory network (AT-BLSTM) model, which consists of bidirectional long short-term memory (BLSTM) layer and attention layer. The BLSTM layer contains forward and backward long short-term memory (LSTM), which uses a novel method to combine forward and backward output. The attention layer utilizes novel self-attention algorithm to assign different weights according to the correlation between features. In this case, AT-BLSTM can predict travel order quantity more accurately than other time series models. For example, AT-BLSTM achieves the lowest MAE (0.0647), the lowest RMSE (0.0836) and the lowest MAPE (0.1239) among all the methods on dataset Xiuying. Extensive experiments on real-world travel datasets offer evidence that the proposed approach matches or outperforms state-of-the-art methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Notes

  1. https://outreach.didichuxing.com/research/opendata/.

References

  1. Malvin N, Pudjaprasetya SR (2021) Staggered conservative scheme for simulating the emergence of a Jamiton in a phantom traffic jam. Int J Intell Transp Syst Res 19(1):128–140. https://doi.org/10.1007/s13177-020-00229-y

    Article  Google Scholar 

  2. Yin X, Wu G, Wei J, Shen Y, Qi H, Yin B (2021) Multi-stage attention spatial-temporal graph networks for traffic prediction. Neurocomputing 428:42–53. https://doi.org/10.1016/j.neucom.2020.11.038

    Article  Google Scholar 

  3. Bai L, Yao L, Kanhere SS, Yang Z, Chu J, Wang X (2019a) Passenger demand forecasting with multi-task convolutional recurrent neural networks. Adv Knowl Discov Data Min 11440:29–42. https://doi.org/10.1007/978-3-030-16145-3_3

    Article  Google Scholar 

  4. Bai L, Yao L, Kanhere SS, Wang X, Sheng QZ (2019b) Passenger demand forecasting with multi-task convolutional recurrent neural networks. Adv Knowl Discov Data Min 11440:29–42. https://doi.org/10.1007/978-3-030-16145-3_3

    Article  Google Scholar 

  5. Zhang J, Zheng Y, Qi D, Li R, Yi X (2016) STG2Seq: spatial-temporal graph to sequence model for multi-step passenger demand forecasting. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Macao, China, pp 1981–1987. https://doi.org/10.24963/ijcai.2019/274

  6. Zhang J, Zheng Y, Qi D (2017) DNN-based prediction model for spatio-temporal data. In: Proceedings of the 24th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems. ACM, Burlingame, California, pp 1–4. https://doi.org/10.1145/2996913.2997016

  7. Zhang K, Feng Z, Chen S, Huang K, Wang G (2016) Deep spatio-temporal residual networks for citywide crowd flows prediction. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 1655–1661

  8. Wang D, Cao W, Xu M, Li J (2016) ETCPS: an effective and scalable traffic condition prediction system. In: International Conference on Database Systems for Advanced Applications. Springer International Publishing, Cham, vol 9643, pp 419–436. https://doi.org/10.1007/978-3-319-32049-6_26

  9. Li X, Pan G, Wu Z, Qi G, Li S, Zhang D, Zhang W, Wang Z (2012) Prediction of urban human mobility using large-scale taxi traces and its applications. Front Comput Sci 6(1):111–121. https://doi.org/10.1007/s11704-011-1192-6

    Article  MathSciNet  Google Scholar 

  10. Moreira-Matias L, Gama J, Ferreira M, Mendes-Moreira J, Damas L (2013) Predicting taxi-passenger demand using streaming data. IEEE Trans Intell Transp Syst 14(3):1393–1402. https://doi.org/10.1109/TITS.2013.2262376

    Article  Google Scholar 

  11. Zhou X, Shen Y, Zhu Y, Huang L (2018) Predicting multi-step citywide passenger demands using attention-based neural networks. In: Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining. ACM, Marina Del Rey, pp 736–744. https://doi.org/10.1145/3159652.3159682

  12. Huang W, Song G, Hong H, Xie K (2014) Deep architecture for traffic flow prediction: deep belief networks with multitask learning. IEEE Trans Intell Transp Syst 15(5):2191–2201

    Article  Google Scholar 

  13. Zhao N, Ye Z, Pei Y, Liang YC, Niyato D (2020) Spatial-temporal attention-convolution network for citywide cellular traffic prediction. IEEE Commun Lett 24(11):5

    Article  Google Scholar 

  14. Poole MA, OÃFarrell PN (1971) The assumptions of the linear regression model. Trans Inst Br Geogr 52:145–158

    Article  Google Scholar 

  15. Ogutu JO, Schulz-Streeck T, Piepho HP (2012) Genomic selection using regularized linear regression models: ridge regression, lasso, elastic net and their extensions. BMC Proc 6(S2):S10. https://doi.org/10.1186/1753-6561-6-S2-S10

    Article  Google Scholar 

  16. Granger C, Andersen A (1978) On the invertibility of time series models. Stoch Process Appl 8(1):87–92. https://doi.org/10.1016/0304-4149(78)90069-8

    Article  MathSciNet  MATH  Google Scholar 

  17. Asteriou D, Hall SG (2016) ARIMA models and the Box–Jenkins methodology. Appl Econ. https://doi.org/10.1057/978-1-137-41547-9_13

    Article  Google Scholar 

  18. Li Y, Zheng Y, Zhang H, Chen L (2015) Traffic prediction in a bike-sharing system. In: Proceedings of the 23rd SIGSPATIAL International Conference on Advances in Geographic Information Systems. ACM, Seattle, pp 1–10. https://doi.org/10.1145/2820783.2820837

  19. Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3th International Conference on Learning Representations

  20. Saon G, Tuske Z, Bolanos D, Kingsbury B (2021) Advancing RNN transducer technology for speech recognition. In: ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Toronto, pp 5654–5658. https://doi.org/10.1109/ICASSP39728.2021.9414716

  21. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444. https://doi.org/10.1038/nature14539

    Article  Google Scholar 

  22. Elman JL (1991) Distributed representations, simple recurrent networks, and grammatical structure. Mach Learn 7:195–225. https://doi.org/10.1007/bf00114844

    Article  Google Scholar 

  23. Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1(2):270–280. https://doi.org/10.1162/neco.1989.1.2.270

    Article  Google Scholar 

  24. Cho K, van Merrienboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, Doha, pp 1724–1734. https://doi.org/10.3115/v1/D14-1179

  25. Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:141235551412.3555

  26. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735

    Article  Google Scholar 

  27. Ma X, Tao Z, Wang Y, Yu H, Wang Y (2015) Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp Res Part C Emerg Technol 54:187–197. https://doi.org/10.1016/j.trc.2015.03.014

    Article  Google Scholar 

  28. Yu R, Li Y, Shahabi C, Demiryurek U, Liu Y (2017) Deep learning: a generic approach for extreme condition traffic forecasting. In: Proceedings of the 2017 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics, Philadelphia, pp 777–785. https://doi.org/10.1137/1.9781611974973

  29. Yao H, Wu F, Ke J, Tang X, Jia Y, Lu S, Gong P, Ye J, Li Z (2018) Deep multi-view spatial-temporal network for taxi demand prediction. Proc AAAI Conf Artif Intell 32:2588–2595

    Google Scholar 

  30. Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5(2):157–166. https://doi.org/10.1109/72.279181

    Article  Google Scholar 

  31. Gehring J, Auli M, Grangier D, Dauphin Y (2017) A convolutional encoder model for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Vancouver, pp 123–135. https://doi.org/10.18653/v1/P17-1012

  32. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30:5998–6008

    Google Scholar 

  33. Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y (2018) Graph attention networks. In: 6th International Conference on Learning Representations

  34. Cheng J, Dong L, Lapata M (2016) Long short-term memory-networks for machine reading. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/D16-1053

  35. Zhou P, Shi W, Tian J, Qi Z, Li B, Hao H, Xu B (2016) Attention-based bidirectional long short-term memory networks for relation classification. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, Berlin, pp 207–212. https://doi.org/10.18653/v1/P16-2034

  36. Lin Z, Feng M, dos Santos CN, Yu M, Xiang B, Zhou B, Bengio Y (2017) A structured self-attentive sentence embedding. In: 5th International Conference on Learning Representations

  37. Kipf TN, Welling M (2017) Semi-supervised classification with graph convolutional networks. In: 5th International Conference on Learning Representations

  38. Qin Y, Song D, Chen H, Cheng W, Jiang G, Cottrell G (2017) A dual-stage attention-based recurrent neural network for time series prediction. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, pp 2627–2633. https://doi.org/10.24963/ijcai.2017/366

  39. Huang L, Ma Y, Wang S, Liu Y (2019) An attention-based spatiotemporal LSTM network for next POI recommendation. IEEE Trans Serv Comput. https://doi.org/10.1109/TSC.2019.2918310

    Article  Google Scholar 

  40. Fernando T, Denman S, Sridharan S, Fookes C (2018) Soft + hardwired attention: an LSTM framework for human trajectory prediction and abnormal event detection. Neural Netw 108:466–478

    Article  Google Scholar 

  41. Liu Q, Wang B, Zhu Y (2020) Short-term traffic speed forecasting based on attention convolutional neural network for arterials. Comput Aided Civ Infrastruct Eng 33(11):999–1016. https://doi.org/10.1111/mice.12417

    Article  Google Scholar 

  42. Chung JH, Kim DW, Kang TK, Lim MT (2020) Traffic sign recognition in harsh environment using attention based convolutional pooling neural network. Neural Process Lett 51(3):2551–2573. https://doi.org/10.1007/s11063-020-10211-0

    Article  Google Scholar 

  43. Yang T, Nan Z, Zhang H, Chen S, Zheng N (2020) Traffic agent trajectory prediction using social convolution and attention mechanism. In: 2020 IEEE Intelligent Vehicles Symposium (IV), pp 278–283. https://doi.org/10.1109/IV47402.2020.9304645

  44. Guo G, Yuan W (2020) Short-term traffic speed forecasting based on graph attention temporal convolutional networks. Neurocomputing 410:387–393. https://doi.org/10.1016/j.neucom.2020.06.001

    Article  Google Scholar 

  45. Xie Y, Xiong Y, Zhu Y (2020) SAST-GNN: a self-attention based spatio-temporal graph neural network for traffic prediction. In: Database Systems for Advanced Applications. Springer International Publishing, Cham, pp 707–714

  46. Fang S, Zhang Q, Meng G, Xiang S, Pan C (2019) GSTNet: global spatial-temporal network for traffic flow prediction. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Macao, China, pp 2286–2293. https://doi.org/10.24963/ijcai.2019/317

  47. Chang B, Park Y, Park D, Kim S, Kang J (2018) Content-aware hierarchical point-of-interest embedding model for successive POI recommendation. In: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Stockholm, Sweden, pp 3301–3307. https://doi.org/10.24963/ijcai.2018/458

  48. Zhao P, Zhu H, Liu Y, Xu J, Li Z, Sheng VS, Zhou X (2020) Where to go next: a spatio-temporal gated network for next POI recommendation. IEEE Trans Knowl Data Eng

  49. Wang D, Cao W, Li J, Ye J (2017) DeepSD: supply-demand prediction for online car-hailing services using deep neural networks. In: 2017 IEEE 33rd International Conference on Data Engineering (ICDE). IEEE, San Diego, pp 243–254. https://doi.org/10.1109/ICDE.2017.83

  50. Lv Y, Duan Y, Kang W, Li Z, Wang FY (2014) Traffic flow prediction with big data: a deep learning approach. IEEE Trans Intell Transp Syst 16(2):865–873. https://doi.org/10.1109/TITS.2014.2345663

    Article  Google Scholar 

  51. Duan Y, LV Y, Wang FY (2016) Travel time prediction with LSTM neural network. In: 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC). IEEE, Rio de Janeiro, pp 1053–1058. https://doi.org/10.1109/ITSC.2016.7795686

  52. Ran X, Shan Z, Fang Y, Lin C (2019) An LSTM-based method with attention mechanism for travel time prediction. Sensors 19(4):861. https://doi.org/10.3390/s19040861

    Article  Google Scholar 

  53. Petersen NC, Rodrigues F, Pereira FC (2019) Multi-output bus travel time prediction with convolutional LSTM neural network. Expert Syst Appl 120:426–435. https://doi.org/10.1016/j.eswa.2018.11.028

    Article  Google Scholar 

  54. Zhang J, Wen Y, Long M, Wang J, Wang H (2020) Predictive recurrent networks for seasonal spatiotemporal data with applications to urban computing. Chin J Comput 43(02):286–302

    Google Scholar 

  55. Yan X, Mou L, Li G, Chen Y, Peng H, Jin Z (2015) Classifying relations via long short term memory networks along shortest dependency path. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp 1785–1794

  56. Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: 3th International Conference on Learning Representations

  57. Hoang ND, Nguyen QL (2019) A novel method for asphalt pavement crack classification based on image processing and machine learning. Eng Comput 35(2):487–498. https://doi.org/10.1007/s00366-018-0611-9

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China (Grant No. 61772386), and the Open Fund of Key Laboratory of Urban Land Resources Monitoring and Simulation, Ministry of Natural Resources (Grant No. KF-2020-05-014).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Huyin Zhang.

Ethics declarations

Conflict of interest

We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service and/or company that could be construed as influencing that position in, or the review of, the manuscript entitled.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, F., Zhang, H. & Tao, S. Travel order quantity prediction via attention-based bidirectional LSTM networks. J Supercomput 78, 4398–4420 (2022). https://doi.org/10.1007/s11227-021-04032-8

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-021-04032-8

Keywords

Navigation