Skip to main content

Research on Multi-Modal Time Series Data Prediction Method Based on Dual-Stage Attention Mechanism

  • Conference paper
  • First Online:
Data Science (ICPCSEE 2023)

Abstract

The production data in the industrial field have the characteristics of multimodality, high dimensionality and large correlation differences between attributes. Existing data prediction methods cannot effectively capture time series and modal features, which leads to prediction hysteresis and poor prediction stability. Aiming at the above problems, this paper proposes a time-series and modal feature enhancement method based on a dual-stage self-attention mechanism (DATT), and a time series prediction method based on a gated feedforward recurrent unit (GFRU). On this basis, the DATT-GFRU neural network with a gated feedforward recurrent neural network and dual-stage self-attention mechanism is designed and implemented. Experiments show that the prediction effect of the neural network prediction model based on DATT is significantly improved. Compared with the traditional prediction model, the DATT-GFRU neural network has a smaller average error of model prediction results, stable prediction performance, and strong generalization ability on the three datasets with different numbers of attributes and different training sample sizes.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Sun, Q.S., Zhang, J.X., Cheng, H.Y., Zhang, Q., Wei, X.P.: Financial time series data prediction by attention-based convolutional neural network. J. Comput. Appl. 42(S2), 290–295 (2022)

    Google Scholar 

  2. Wang, T., Wang, M.: Communication network time series prediction algorithm based on big data method. Wireless Pers. Commun. 102(2), 1041–1056 (2017). https://doi.org/10.1007/s11277-017-5138-7

    Article  Google Scholar 

  3. Li, H.L.: Dynamic time warping based on time weighting for time series data mining. Inform. Sci. 547 (2021)

    Google Scholar 

  4. Li, J., Zhu, L., Zhang, Y., Guo, D., Xia, X.: Attention-based multi-scale prediction network for time-series data. China Commun. 19(5), 286–301 (2022)

    Article  Google Scholar 

  5. Wang, H., Zhang, Z.: TATCN: Time series prediction model based on time attention mechanism and TCN. In: 2022 IEEE 2nd International Conference on Computer Communication and Artificial Intelligence (CCAI), pp. 26–31. Beijing, China (2022)

    Google Scholar 

  6. Xing, Z.K., He, Y.G.: Multi-modal information analysis for fault diagnosis with time-series data from power transformer. Int. J. Electric. Power Energy Syst. 144 (2023)

    Google Scholar 

  7. Elman, J.L.: Finding structure in time. Cogn. Sci. 14(2), 179–211 (1990)

    Google Scholar 

  8. Heimes, F.O.: Recurrent neural networks for remaining useful life estimation. Prognostics and Health Management, 2008. PHM 2008. International Conference on IEEE (2008)

    Google Scholar 

  9. Zhao, J., Zeng, D., Liang, S., et al.: Prediction model for stock price trend based on recurrent neural network. J. Ambient Intell. Hum. Comput. 12, 745–753 (2021)

    Article  Google Scholar 

  10. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  11. Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM networks. In: 2005 IEEE International Joint Conference on Neural Networks, 2005. IEEE 4, 2047–2052 (2005)

    Google Scholar 

  12. Eck, D., Graves, A., Schmidhuber, J.: A new approach to continuous speech recognition using LSTM recurrent neural networks. Technical Report (2003)

    Google Scholar 

  13. Shi, X., Chen, Z., Wang, H., et al.: Convolutional LSTM Network. A Machine Learning Approach for Precipitation Nowcasting (2015)

    Google Scholar 

  14. Alahi, A., Goel, K., Ramanathan, V., et al.: Social LSTM: human trajectory prediction in crowded spaces. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 961–971 (2016)

    Google Scholar 

  15. Tong, Y., Tien, I.: Time-series prediction in nodal networks using recurrent neural networks and a pairwise-gated recurrent unit approach. ASCE-ASME J. Risk Uncert. Eng. Syst. Part A. Civil Eng. 8(2), 04022002 (2022)

    Article  Google Scholar 

  16. Cho, K., Van Merriënboer, B., Gulcehre, C., et al.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv: 1406.1078 (2014)

    Google Scholar 

  17. Zhang, D., Kabuka, M.R.: Combining weather condition data to predict traffic flow: a GRU-based deep learning approach. IET Intel. Transport Syst. 12(7), 578–585 (2018)

    Article  Google Scholar 

  18. Aiken, E. L., Nguyen, A. T., Santillana, M.: Towards the Use of Neural Networks for Influenza Prediction at Multiple Spatial Resolutions. arXiv preprint arXiv:1911.02673 (2019)

  19. Sun, P., Boukerche, A., Tao, Y.: SSGRU: a novel hybrid stacked GRU-based traffic volume prediction approach in a road network. Comput. Commun. 160, 502–511 (2020)

    Article  Google Scholar 

  20. Ji, S.P., Meng, Y.L., Yan, L., et al.: GRU-corr neural network optimized by improved PSO algorithm for time series prediction. Int. J. Artific. Intell. Tools 29(07n08), 2040010 (2020)

    Google Scholar 

  21. Gao, C., Zhang, N., Li, Y., et al.: Self-attention-based time-variant neural networks for multi-step time series forecasting. Neural Comput. Appl. 34, 8737–8754 (2022)

    Article  Google Scholar 

  22. Su, Y., Cui, C., Qu, H.: Self-attentive moving average for time series prediction. Appl. Sci. 12(7), 3602 (2022)

    Article  Google Scholar 

  23. Noh, S.H.: Analysis of gradient vanishing of RNNs and performance comparison. Information 12(11), 442 (2021)

    Article  Google Scholar 

  24. Frame, J.M., Kratzert, F., Raney, A., et al.: Post-processing the national water model with long short-term memory networks for streamflow predictions and model diagnostics. JAWRA J. Am. Water Resourc. Assoc. 57(6), 885–905 (2021)

    Article  Google Scholar 

  25. Zeng, C., Ma, C., Wang, K., Cui, Z.: Parking occupancy prediction method based on multi factors and stacked GRU-LSTM. IEEE Access 10, 47361–47370 (2022)

    Article  Google Scholar 

Download references

Acknowledgements

This work is financially supported by: The National Key R&D Program of China (No. 2020YFB1712600); The Fundamental Research Funds for Central University (No. 3072022QBZ0601); The National Natural Science Foundation of China (No. 62272126); and The National Natural Science Foundation of China (No. 61872104).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yulong Meng .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, X. et al. (2023). Research on Multi-Modal Time Series Data Prediction Method Based on Dual-Stage Attention Mechanism. In: Yu, Z., et al. Data Science. ICPCSEE 2023. Communications in Computer and Information Science, vol 1879. Springer, Singapore. https://doi.org/10.1007/978-981-99-5968-6_10

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-5968-6_10

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-5967-9

  • Online ISBN: 978-981-99-5968-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics