Skip to main content
Log in

Indoor localization system using deep learning based scene recognition

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

One of the major problems of indoor positioning using Wi-Fi is the unstable wireless signal in a large area inside the building. In some buildings, signal path loss occurs due to the partition separating the space. In this case, since it is difficult to identify a specific partition when indoor positioning is performed using only received signal strength indicator (RSSI), a partitioning or segmentation algorithm is required to solve this problem. The proposed system in this paper consists of a scene recognition algorithm suitable for a given environment, an image-based indoor location awareness algorithm (IILAA), and a clustering algorithm that connects these modules to identify not only the exact spatial location but also the scene cluster of the user. The system is composed of a scene recognition module which identifies an image taken by the user. A convolutional neural network trained on resnet50 architecture will identify the scene and forward the classified scene result to another module name mapping algorithm module where KNN is used to process the RSSI received by the user mobile device. As a result of the experiment, first, it was confirmed that the proposed system can recognize an average of 93.07% of a given indoor environment. In addition, it was confirmed that the average error distance of the proposed system is 1.31 m, which is about 6.2 times more accurate than 8.15 m, which is the average error distance of fingerprinting using RSSI. The main factor of this performance result is that the proposed system appropriately selects the classified cluster for each scene in the indoor space. Second, it was confirmed that the proposed IILAA’s error distance performance was 1.75 m in an environment in which both clustering/non-clustering fingerprint maps were combined, which is about 3.90 times more accurate than the 6.90 m average error distance of conventional fingerprinting. Finally, the trained model was used on MIT67 dataset to compare its 87. 6% accuracy with other approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19

Similar content being viewed by others

References

  1. Bansal M, Kumar M, Kumar M (2021) 2D object recognition techniques: state-of-the-art work. Archives of Computational Methods in Engineering 28(3):1147–1161

    Article  MathSciNet  Google Scholar 

  2. Bansal M, Kumar M, Kumar M, Kumar K (2021) Content-based image retrieval system using ORB and SIFT features. Neural Comput Appl 32(7):2725–2733

    Google Scholar 

  3. Chen BX, Sahdev R, Wu D, Zhao X, Papagelis M, Tsotsos JK (2019) Scene Classification inIndoor Environments for Robots using Context Based Word Embeddings. IEEEInternational Conference of Robotics and Automation (ICRA) Workshop (ICRA) Workshop

  4. Cheng X, Lu J, Feng J, Yuan B, Zhou J (2018) Scene recognition with objectness. Pattern Recogn 74(C):474–487

  5. Chhabra P, Garg NK, Kumar M (2018) An efficient content-based image retrieval system using BayesNet and K-NN. Multimed Tools Appl 77(16):21557–21570

    Article  Google Scholar 

  6. Chhabra P, Garg NK, Kumar M (2020) An efficient technique for object recognition using Shi-Tomasi corner detection algorithm. Soft Comput 25(6):4423–4432

    Google Scholar 

  7. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference, vol 1. IEEE, pp 886–893

  8. Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp 1440–1448

  9. Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580–587

  10. Gong Y, Wang L, Guo R, Lazebnik S (2014)Multi-scale orderless pooling of deep convolutional activation features. In: Proc Eur Conf Comput Vis (ECCV), pp 392–407

  11. Guo Y, Liu Y, Lao S, Bakker EM, Bai L, Lew MS (2018) Bag of surrogate parts feature for visual recognition. IEEE Trans Multimed 20(6):1525–1536

    Article  Google Scholar 

  12. Güzel MS (2013) Autonomous vehicle navigation using vision and mapless strategies: a survey. Advances in Mechanical Engineering. https://doi.org/10.1155/2013/234747

  13. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

  14. Huang G, Liu Z, Maaten LVD, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708

  15. Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally KWJ, Keutzer K (2016) SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and < 0.5MB model size. arXiv preprintarXiv:1602.07360

  16. Labinghisa BA (2018) A study on predictive path learning based localization algorithm in indoor crowdsourced Wi-Fi environment. Thesis for Degree of Master of Engineering, The Graduate School of Tongmyong University, pp 1–58

  17. Labinghisa BA (2020) A study of indoor localization system using reliable deep learning based scene recognition. Thesis for Degree of Doctor of Engineering, The Graduate School of Tongmyong University, pp 1–87

  18. Labinghisa BA, Lee DM (2021) Neural network based indoor localization system with enhanced virtual access points. J Supercomput (SUPE) 7(1):(ISSN 0920 – 8542), Springer, Berlin, pp 638–651

  19. Lin TY, RoyChowdhury A, Maji S (2018) Bilinear convolutional neural networks for fine-grained visual recognition. IEEE Trans Pattern Anal Mach Intell 40(6):1309–1322

    Article  Google Scholar 

  20. Liu H, Darabi H, Banerjee P, Liu J (2007) Survey of wireless indoor positioning techniques and systems. IEEE Trans Syst Man Cybern C Appl Rev 37(6):1067–1080

    Article  Google Scholar 

  21. Machaj J, Brida P, Tatarova B (2010) Impact of the number of access points in indoor fingerprinting localization. In: 20th International Conference Radioelektronika, pp 1–4

  22. Matei A, Glavan A, Talavera E, Martínez (2020) Deep learning for scene recognition from visual data: a survey. ArXiv abs/2007.01806

  23. Quattoni A, Torralba A (2009) Recognizing indoor scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 413–420

  24. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520

  25. Tang P, Wang H, Kwong S (2017) G-MS2F: GoogLeNet based multistage feature fusion of deep CNN for scene recognition. Neurocomputing 225:188–197

    Article  Google Scholar 

  26. Uijlings JRR, van de Sande KEA, Gevers T, Smeulders AWM (2013) Selective search for object recognition. Int J Comput Vis Vol 104:154–171

    Article  Google Scholar 

  27. Walther D, Itti L, Riesenhuber M, Poggio T, Koch C (2002) Attentional selection for object recognition - a gentle way. In: International Workshop on Biologically Motivated Computer Vision. Springer, Berlin, pp 472–479

  28. Wang Z, Wang L, Wang Y, Zhang B, Qiao Y (2017) Weakly supervised patchnets: describing and aggregating local patches for scene recognition. IEEE Trans Image Process 26(4):2028–2041

    Article  MathSciNet  Google Scholar 

  29. Zhou B, Lapedriza A, Khosla A, Oliva A, Torralba A (2017) Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, pp 1452–1464

Download references

Acknowledgements

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2019R1F1A1062670).

This research was supported by the BB21plus funded by Busan Metropolitan City and Busan Institute for Talent & Lifelong Education (BIT).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dong Myung Lee.

Ethics declarations

Conflict of interest/Competing interests

The authors declare that there are no conflicts of interests/competing interests regarding the publication of this article.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Labinghisa, B.A., Lee, D.M. Indoor localization system using deep learning based scene recognition. Multimed Tools Appl 81, 28405–28429 (2022). https://doi.org/10.1007/s11042-022-12481-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-022-12481-3

Keywords

Navigation