Abstract
Indoor localisation and tracking in a residential home setting are envisaged to play an essential role in smart home environments. As it is hard to obtain a large number of labelled data, semi-supervised learning with Reinforcement Learning is considered in this paper. We extend the Reinforcement Learning approach, and propose a reward function that provides a clear interpretation and defines an objective function of the Reinforcement Learning. Our interpretable reward allows us to extend the model to incorporate multiple sources of information. We also provide a connection between our approach and a conventional inference algorithm for Conditional Random Field, Hidden Markov Model and Maximum Entropy Markov Model. The developed framework shows that our approach benefits over the conventional algorithms a real-time prediction scenarios. The proposed Reinforcement Learning method is compared against other supervised learning approaches. The results suggest that our method can learn in the semi-supervised learning setting and performs well in a small labelled data regime.
This work was supported by SPHERE Next Steps Project funded by the U.K. Engineering, and Physical Sciences Research Council (EPSRC) under Grant EP/R005273/1 and the UKRI Turing AI Fellowship EP/V024817/1.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006)
Hähnel, B.F.D., Fox, D.: Gaussian processes for signal strength-based location estimation. In: Proceeding of Robotics: Science and Systems (2006)
Kozlowski, M., Byrne, D., Santos-Rodriguez, R., Piechocki, R.: Data fusion for robust indoor localisation in digital health. In: IEEE WCNCW (2018)
Kozlowski, M., Santos-Rodriguez, R., Piechocki, R.J.: Sensor modalities and fusion for robust indoor localisation. EAI Endorsed Trans. Ambient Syst. 6(18), e5 (2019)
Lafferty, J.D., McCallum, A., Pereira, F.C.N.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: Proceedings of the Eighteenth International Conference on Machine Learning. ICML 2001, San Francisco, CA, USA, pp. 282–289. Morgan Kaufmann Publishers Inc. (2001)
Levine, S.: Reinforcement learning and control as probabilistic inference: tutorial and review. arXiv preprint arXiv:1805.00909 (2018)
Li, Y., Hu, X., Zhuang, Y., Gao, Z., Zhang, P., El-Sheimy, N.: Deep reinforcement learning: another perspective for unsupervised wireless localization. IEEE Internet Things J. 7, 6279–6287 (2019)
McCallum, A., Freitag, D., Pereira, F.C.N.: Maximum entropy Markov models for information extraction and segmentation. In: ICML (2000)
Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529 (2015)
Mohammadi, M., Al-Fuqaha, A., Guizani, M., Oh, J.S.: Semisupervised deep reinforcement learning in support of IoT and smart city services. IEEE Internet Things J. 5, 624–635 (2017)
Sun, S., Li, Y., Rowe, W., Wang, X., Kealy, A., Moran, B.: Practical evaluation of a crowdsourcing indoor localization system using hidden Markov models. IEEE Sens.s J. 19, 9332–9340 (2019)
Twomey, N., et al.: The SPHERE challenge: activity recognition with multimodal sensor data. arXiv:1603.00797 (2016)
Zhang, W., Liu, K., Zhang, W., Zhang, Y., Gu, J.: Deep neural networks for wireless localization in indoor and outdoor environments. Neurocomputing 194, 279–287 (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Yamagata, T., Santos-Rodríguez, R., Piechocki, R., Flach, P. (2022). Understanding Reinforcement Learning Based Localisation as a Probabilistic Inference Algorithm. In: Pimenidis, E., Angelov, P., Jayne, C., Papaleonidas, A., Aydin, M. (eds) Artificial Neural Networks and Machine Learning – ICANN 2022. ICANN 2022. Lecture Notes in Computer Science, vol 13530. Springer, Cham. https://doi.org/10.1007/978-3-031-15931-2_10
Download citation
DOI: https://doi.org/10.1007/978-3-031-15931-2_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-15930-5
Online ISBN: 978-3-031-15931-2
eBook Packages: Computer ScienceComputer Science (R0)