Skip to main content

Understanding Reinforcement Learning Based Localisation as a Probabilistic Inference Algorithm

  • Conference paper
  • First Online:
Artificial Neural Networks and Machine Learning – ICANN 2022 (ICANN 2022)

Abstract

Indoor localisation and tracking in a residential home setting are envisaged to play an essential role in smart home environments. As it is hard to obtain a large number of labelled data, semi-supervised learning with Reinforcement Learning is considered in this paper. We extend the Reinforcement Learning approach, and propose a reward function that provides a clear interpretation and defines an objective function of the Reinforcement Learning. Our interpretable reward allows us to extend the model to incorporate multiple sources of information. We also provide a connection between our approach and a conventional inference algorithm for Conditional Random Field, Hidden Markov Model and Maximum Entropy Markov Model. The developed framework shows that our approach benefits over the conventional algorithms a real-time prediction scenarios. The proposed Reinforcement Learning method is compared against other supervised learning approaches. The results suggest that our method can learn in the semi-supervised learning setting and performs well in a small labelled data regime.

This work was supported by SPHERE Next Steps Project funded by the U.K. Engineering, and Physical Sciences Research Council (EPSRC) under Grant EP/R005273/1 and the UKRI Turing AI Fellowship EP/V024817/1.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006)

    MATH  Google Scholar 

  2. Hähnel, B.F.D., Fox, D.: Gaussian processes for signal strength-based location estimation. In: Proceeding of Robotics: Science and Systems (2006)

    Google Scholar 

  3. Kozlowski, M., Byrne, D., Santos-Rodriguez, R., Piechocki, R.: Data fusion for robust indoor localisation in digital health. In: IEEE WCNCW (2018)

    Google Scholar 

  4. Kozlowski, M., Santos-Rodriguez, R., Piechocki, R.J.: Sensor modalities and fusion for robust indoor localisation. EAI Endorsed Trans. Ambient Syst. 6(18), e5 (2019)

    Google Scholar 

  5. Lafferty, J.D., McCallum, A., Pereira, F.C.N.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: Proceedings of the Eighteenth International Conference on Machine Learning. ICML 2001, San Francisco, CA, USA, pp. 282–289. Morgan Kaufmann Publishers Inc. (2001)

    Google Scholar 

  6. Levine, S.: Reinforcement learning and control as probabilistic inference: tutorial and review. arXiv preprint arXiv:1805.00909 (2018)

  7. Li, Y., Hu, X., Zhuang, Y., Gao, Z., Zhang, P., El-Sheimy, N.: Deep reinforcement learning: another perspective for unsupervised wireless localization. IEEE Internet Things J. 7, 6279–6287 (2019)

    Article  Google Scholar 

  8. McCallum, A., Freitag, D., Pereira, F.C.N.: Maximum entropy Markov models for information extraction and segmentation. In: ICML (2000)

    Google Scholar 

  9. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529 (2015)

    Article  Google Scholar 

  10. Mohammadi, M., Al-Fuqaha, A., Guizani, M., Oh, J.S.: Semisupervised deep reinforcement learning in support of IoT and smart city services. IEEE Internet Things J. 5, 624–635 (2017)

    Article  Google Scholar 

  11. Sun, S., Li, Y., Rowe, W., Wang, X., Kealy, A., Moran, B.: Practical evaluation of a crowdsourcing indoor localization system using hidden Markov models. IEEE Sens.s J. 19, 9332–9340 (2019)

    Article  Google Scholar 

  12. Twomey, N., et al.: The SPHERE challenge: activity recognition with multimodal sensor data. arXiv:1603.00797 (2016)

  13. Zhang, W., Liu, K., Zhang, W., Zhang, Y., Gu, J.: Deep neural networks for wireless localization in indoor and outdoor environments. Neurocomputing 194, 279–287 (2016)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Taku Yamagata .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yamagata, T., Santos-Rodríguez, R., Piechocki, R., Flach, P. (2022). Understanding Reinforcement Learning Based Localisation as a Probabilistic Inference Algorithm. In: Pimenidis, E., Angelov, P., Jayne, C., Papaleonidas, A., Aydin, M. (eds) Artificial Neural Networks and Machine Learning – ICANN 2022. ICANN 2022. Lecture Notes in Computer Science, vol 13530. Springer, Cham. https://doi.org/10.1007/978-3-031-15931-2_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-15931-2_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-15930-5

  • Online ISBN: 978-3-031-15931-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics