Abstract
Some of the limitations of state-space models are given by the difficulty of modelling certain systems, the filters convergence time or the impossibility of modelling dependencies in the long term. Having agile and alternative methodologies that allow the modelling of complex problems but still provide solutions to the classic challenges of estimation or filtering, such as the position estimation of a mobile with noisy measurements of the same variable, are of high interest. In this work, we address the problem of position estimation of 1-D dynamic systems from a deep learning paradigm, using Long-Short Term Memory (LSTM) architectures designed to solve problems with long term temporal dependencies, in combination with other recurrent networks. A deep neuronal architecture inspired by the Encoder-Decoder language systems is implemented, remarking its limits and finding a solution capable of making position estimations of a moving object. The results are finally compared with the optimal values from the Kalman filter, obtaining comparable results in error terms.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Rudy, S.H., Kutz, J.N., Brunton, S.L.: Deep learning of dynamics and signal-noise decomposition with time-stepping constraints. J. Comput. Phys. 396, 483–506 (2019)
Afshari, H.H., Gadsden, S.A., Habibi, S.: Gaussian filters for parameter and state estimation: a general review of theory and recent trends. Sig. Process. 135, 218–238 (2017)
Ali, J.M., Hussain, M.A., Tade, M.O., Zhang, J.: Artificial intelligence techniques applied as estimator in chemical process systems - a literature survey. Expert Syst. Appl. 42(14), 5915–5931 (2015)
Hochreiter, S., Schmidhuber, U.J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Bahdanau, D., Cho, K.H., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: 3rd International Conference on Learning Representation. ICLR 2015 - Conference Track Proceedings, pp. 1–15 (2015)
Gan, C., Wang, L., Zhang, Z., Wang, Z.: Sparse attention based separable dilated convolutional neural network for targeted sentiment analysis. Knowl.-Based Syst. 188, 104827 (2019)
Wang, Y., Huang, M., Zhao, L., Zhu, X.: Attention-based LSTM for aspect-level sentiment classification. In: Proceeding Conference on Empirical Methods in Natural Language Processing. EMNLP 2016, pp. 606–615 (2016)
Arriaga, O., Plöger, P., Valdenegro-Toro, M.: Image captioning and classification of dangerous situations no. 1 (2017)
Arsene, C.T.C., Hankins, R., Yin, H.: Deep learning models for denoising ECG signals. In: 2019 European Signal Processing Conference, 2–6 September, vol. 2017, no. Iaa 220, pp. 1–5 (2019)
Song, X., et al.: Time-series well performance prediction based on long short-term memory (LSTM) neural network model. J. Pet. Sci. Eng. 186, 106682 (2019)
Zhao, Z., Chen, W., Wu, X., Chen, P.C.V., Liu, J.: LSTM network: a deep learning approach for short-term traffic forecast. IET Image Process. 11(1), 68–75 (2017)
Orimoloye, L.O., Sung, M.C., Ma, T., Johnson, J.E.V.: Comparing the effectiveness of deep feedforward neural networks and shallow architectures for predicting stock price indices. Expert Syst. Appl. 139, 112828 (2020)
Zaheer, M., Ahmed, A., Smola, A.J.: Latent LSTM allocation joint clustering and non-linear dynamic modeling of sequential data. In: 34th International Conference on Machine Learning. ICML 2017, vol. 8, pp. 6040–6049 (2017)
Raissi, M., Perdikaris, P., Karniadakis, G.E.: Multistep neural networks for data-driven discovery of nonlinear dynamical systems, pp. 1–19 (2018)
Zheng, X., Zaheer, M., Ahmed, A., Wang, Y., Xing, E.P., Smola, A.J.L: State space LSTM models with particle MCMC inference, pp. 1–12 (2017)
Shapsough, S., Dhaouadi, R., Zualkernan, I.: Using linear regression and back propagation neural networks to predict performance of soiled PV modules. Procedia Comput. Sci. 155(2018), 463–470 (2019)
Barabanov, N.E., Prokhorov, D.V.: Stability analysis of discrete-time recurrent neural networks. IEEE Trans. Neural Netw. 13(2), 292–303 (2002)
Deng, L., Hajiesmaili, M.H., Chen, M., Zeng, H.: Energy-efficient timely transportation of long-haul heavy-duty trucks. IEEE Trans. Intell. Transp. Syst. 19(7), 2099–2113 (2018)
Wu, Q., Lin, H.: A novel optimal-hybrid model for daily air quality index prediction considering air pollutant factors. Sci. Total Environ. 683, 808–821 (2019)
Kingma, D.P., Ba, J.L.: Adam: a method for stochastic optimization. In: 3rd International Conference on Learning Representation - Conference Track Proceeding. ICLR 2015, pp. 1–15 (2015)
Acknowledgments
This work was supported by Ministry of Science, Innovation and Universities from Spain under grant agreement No. PRE-C-2018-0079.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Llerena, J.P., García, J., Molina, J.M. (2021). An Approach to Forecasting and Filtering Noise in Dynamic Systems Using LSTM Architectures. In: Herrero, Á., Cambra, C., Urda, D., Sedano, J., Quintián, H., Corchado, E. (eds) 15th International Conference on Soft Computing Models in Industrial and Environmental Applications (SOCO 2020). SOCO 2020. Advances in Intelligent Systems and Computing, vol 1268. Springer, Cham. https://doi.org/10.1007/978-3-030-57802-2_15
Download citation
DOI: https://doi.org/10.1007/978-3-030-57802-2_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-57801-5
Online ISBN: 978-3-030-57802-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)