Abstract
Although deep learning makes great success in increasing the accuracy for long sequence time-series forecasting, its complex neural network structure, which comprises many different types of layers and each layer containing hundreds and thousands of neurons, challenges the computing and memory capability of embedded platforms. This paper proposes a lightweight and efficient neural network called TTFNet, which forecasts long time series using three types of features (i.e., the Trend, Temporal attention, and Frequency attention) extracted from raw time series. In TTFNet, we perform a pooling operation on the historical data in a recent time window to extract a general trend, use a multi-layer perceptron to discover the temporal correlation between data as temporal attention, and apply the fast Fourier transforms on data to obtain frequency information as frequency attention. Each feature is separately extracted from its neural network branch with an output result, and we weight the three results to generate the final prediction while optimal weights are learnt. Also, the three prediction results can run in parallel since they are independent from one another. The experimental results show that the proposed method reduces the memory overhead and runtime by 62% and 81% of the five counterpart methods on average while achieving comparable performance.
Similar content being viewed by others
Data availibility
The datasets analyzed during the current study are available in https://github.com/zhouhaoyi/ETDataset, https://archive.ics.uci.edu/ml/datasets/ElectricityLoadDiagrams20112014 and https://www.ncei.noaa.gov/data/local-climatological-data/.
References
Kart U, Lukežič A, Kristan M et al (2019) Object tracking by reconstruction with view-specific discriminative correlation filters. In: 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 1339–1348
Zhang P, Liu W, Wang D et al (2020) Non-rigid object tracking via deep multi-scale spatial-temporal discriminative saliency maps. Pattern Recognit 100:107130
Kalajdjieski J, Korunoski M, Stojkoska BR, et al (2020) Smart city air pollution monitoring and prediction: a case study of skopje. In: ICT Innovations 2020. Machine Learning and Applications, pp 15–27
Xie P, Li T, Liu J et al (2020) Urban flow prediction from spatiotemporal data using machine learning: a survey. Inf Fus 59:1–12
Yang A-M, Han Y, Liu C-S et al (2021) D-tsvr recurrence prediction driven by medical big data in cancer. IEEE Trans Ind Inf 17(5):3508–3517
Ren L, Liu Y, Huang D, Huang K, Yang C (2022) Mctan: a novel multichannel temporal attention-based network for industrial health indicator prediction. IEEE Trans Neural Netw Learn Syst 1–12
Liu F, Xue S, Wu J, et al (2020) Deep learning for community detection: progress, challenges and opportunities. In: Proceedings of the twenty-ninth international joint conference on artificial intelligence, IJCAI-20, pp 4981–4987
Bui T-C, Kim J, Kang T, et al (2021) Star: spatio-temporal prediction of air quality using a multimodal approach. In: Intelligent systems and applications, pp 389–406
Bentsen LØ, Warakagoda ND, Stenbro R et al (2023) Spatio-temporal wind speed forecasting using graph networks and novel transformer architectures. Appl Energy 333:120565
Wu M, Zhu C, Chen L (2020) Multi-task spatial-temporal graph attention network for taxi demand prediction. In: Proceedings of the 2020 5th international conference on mathematics and artificial intelligence. ICMAI 2020, pp 224–228
An J, Guo L, Liu W et al (2021) Igagcn: information geometry and attention-based spatiotemporal graph convolutional networks for traffic flow prediction. Neural Netw 143:355–367
Chaovalit P, Gangopadhyay A, Karabatis G, et al (2011) Discrete wavelet transform-based time series analysis and mining. ACM Comput Surv 43(2)
Mohammadi HA, Ghofrani S, Nikseresht A (2023) Using empirical wavelet transform and high-order fuzzy cognitive maps for time series forecasting. Appl Soft Comput 109990
Ong P, Zainuddin Z (2019) Optimizing wavelet neural networks using modified cuckoo search for multi-step ahead chaotic time series prediction. Appl Soft Comput 80:374–386
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45(11):2673–2681
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: Proceedings of the 27th international conference on neural information processing systems- Volume 2. NIPS’14, pp 3104–3112
Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems NIPS’17, pp 6000–6010
Zhou H, Zhang S, Peng J, et al (2021) Informer: beyond efficient transformer for long sequence time-series forecasting. In: The thirty-fifth AAAI conference on artificial intelligence, AAAI 2021, p
Li S, Jin X, Xuan Y, et al (2019) Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. In: Advances in neural information processing systems, vol. 32
Lee-Thorp J, Ainslie J, Eckstein I, et al (2021) FNet: mixing tokens with fourier transforms, 2105–03824. arXiv:2105.03824 [cs.CL]
Williams BM, Durvasula PK, Brown DE (1998) Urban freeway traffic flow prediction: application of seasonal autoregressive integrated moving average and exponential smoothing models. Transp Res Rec 1644(1):132–141
Kumar SV, Vanajakshi L (2015) Short-term traffic flow prediction using seasonal arima model with limited input data. Eur Transp Res Rev 7(3):1–9
Contreras-Reyes JE (2022) Rényi entropy and divergence for varfima processes based on characteristic and impulse response functions. Chaos Solitons Fractals 160:112268
Karevan Z, Suykens JAK (2020) Transductive LSTM for time-series prediction: an application to weather forecasting. Neural Netw 125:1–9
Yang Y, Fan C, Xiong H (2022) A novel general-purpose hybrid model for time series forecasting. Appl Intell 52(2):2212–2223
Khedhiri S (2022) Comparison of SARFIMA and LSTM methods to model and to forecast Canadian temperature. Region Stat 12(02):177–194
Cho K, van Merriënboer B, Gulcehre C, et al (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1724–1734
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3rd international conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7–9, 2015, conference track proceedings
Li D, Han M, Wang J (2012) Chaotic time series prediction based on a novel robust echo state network. IEEE Trans Neural Netw Learn Syst 23(5):787–799
González-Zapata AM, Tlelo-Cuautle E, Ovilla-Martinez B et al (2022) Optimizing echo state networks for enhancing large prediction horizons of chaotic time series. Mathematics 10(20):3886
Zhang Z, Wang Y, Wang K (2013) Fault diagnosis and prognosis using wavelet packet decomposition, Fourier transform and artificial neural network. J Intell Manuf 24(6):1213–1227
Mironovova M, Bíla J (2015) Fast fourier transform for feature extraction and neural network for classification of electrocardiogram signals. In: 2015 fourth international conference on future generation communication technology (FGCT), pp 1–6
Lin S, Liu N, Nazemi M, et al. (2018) FFT-based deep learning deployment in embedded systems. In: 2018 design, automation test in Europe conference exhibition (DATE), pp 1045–1050
Abtahi T, Shea C, Kulkarni A et al (2018) Accelerating convolutional neural network with FFT on embedded hardware. IEEE Trans.VLSI Syst 26(9):1737–1749
Choromanski K, Likhosherstov V, Dohan D, et al (2020) Masked language modeling for proteins via linearly scalable long-context transformers, 2006–03555. arXiv:2006.03555 [cs.LG]
Sak H, Senior A, Beaufays F (2014) Long short-term memory based recurrent neural network architectures for large vocabulary speech recognition. Comput Sci 338–342
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: Proceedings of the 27th international conference on neural information processing systems - Volume 2. NIPS’14, pp 3104–3112
Acknowledgements
This work is funded in part by the National Natural Science Foundation of China (File No. 62072216) and the Science and Technology Development Fund, Macau SAR (File No. 0076/2022/A2 and 0008/2022/AGJ).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Chen, L., Li, G., Huang, G. et al. A lightweight model using frequency, trend and temporal attention for long sequence time-series prediction. Neural Comput & Applic 35, 21291–21307 (2023). https://doi.org/10.1007/s00521-023-08871-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-023-08871-9