Skip to main content
Log in

DAM-SLAM: depth attention module in a semantic visual SLAM based on objects interaction for dynamic environments

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Nowadays, modern visual SLAM methods attempt to deal with dynamic environments by considering the non-rigid scene assumption. This well-established approach combines geometric and semantic information to detect dynamic objects to achieve accurate localization and mapping in real environments. However, these methods need more generalization and scene awareness because of their reasoning limits due to the labeling object strategy and the need for matched keypoints. Therefore, we propose a novel method called Depth Attention Module SLAM (DAM-SLAM) that overcomes the limitations of existing methods. The main idea is to take into account the depth influence used in the geometric and semantic modules through a depth-related adaptive threshold and impact factor. Moreover, a Bayesian filter is used to refine the keypoints state estimates using a motion probability update function based on a weighting strategy related to the keypoints area (in/out of segmented object’s masks). In addition, we designed a Depth Attention Module that allows generalization to other methods by considering the non-matched keypoints and the keypoints out of segmented regions. This module estimates these keypoints state without requiring any prior semantic information by determining the interactions between the objects. We estimate this interaction through the correlation between the proximity of depth and position of these keypoints with the dynamic keypoints in a specific zone of influence of dynamic objects. The obtained results demonstrate the efficacy of the proposed method in providing accurate localization and mapping in dynamic environments.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Fig. 4
Algorithm 2
Fig. 5
Algorithm 3
Fig. 6
Fig. 7
Fig. 8
Algorithm 4
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Davison AJ, Reid ID, Molton ND, Stasse O (2007) MonoSLAM: Real-time single camera SLAM. IEEE transactions on pattern analysis and machine intelligence 29(6):1052–1067

    Article  Google Scholar 

  2. Klein, G, & Murray, D (2007) Parallel tracking and mapping for small AR workspaces. In 2007 6th IEEE and ACM international symposium on mixed and augmented reality (pp 225-234) IEEE

  3. Mur-Artal R, Montiel JMM, Tardos JD (2015) ORB-SLAM: a versatile and accurate monocular SLAM system. IEEE transactions on robotics 31(5):1147–1163

    Article  Google Scholar 

  4. Campos Carlos et al (2021) Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Transactions on Robotics 37(6):1874–1890

  5. Mur-Artal R, Tardós JD (2017) Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE transactions on robotics 33(5):1255–1262

    Article  Google Scholar 

  6. Qin T, Li P, Shen S (2018) Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Transactions on Robotics 34(4):1004–1020

    Article  Google Scholar 

  7. Bescos B, Fácil JM, Civera J, Neira J (2018) DynaSLAM: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters 3(4):4076–4083

    Article  Google Scholar 

  8. Xiao L, Wang J, Qiu X, Rong Z, Zou X (2019) Dynamic-SLAM: Semantic monocular visual localization and mapping based on deep learning in dynamic environment. Robotics and Autonomous Systems 117:1–16

    Article  Google Scholar 

  9. Yu, C, Liu, Z, Liu, X. J, Xie, F, Yang, Y, Wei, Q, & Fei, Q (2018) DS-SLAM: A semantic visual SLAM towards dynamic environments. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp 1168–1174) IEEE

  10. Yan L, Hu X, Zhao L, Chen Y, Wei P, Xie H (2022) DGS-SLAM: a fast and robust rgbd slam in dynamic environments combined by geometric and semantic information. Remote Sensing 14(3):795

    Article  Google Scholar 

  11. Cheng J, Wang Z, Zhou H, Li L, Yao J (2020) DM-SLAM: A feature-based SLAM system for rigid dynamic scenes. ISPRS International Journal of Geo-Information 9(4):202

    Article  Google Scholar 

  12. Cui L, Ma C (2019) SOF-SLAM: A semantic visual SLAM for dynamic environments. IEEE access 7:166528–166539

    Article  Google Scholar 

  13. Li A, Wang J, Xu M, Chen Z (2021) DP-SLAM: A visual SLAM with moving probability towards dynamic environments. Information Sciences 556:128–142

    Article  Google Scholar 

  14. Liu Y, Miura J (2021) RDMO-SLAM: Real-time visual SLAM for dynamic environments using semantic label prediction with optical flow. IEEE Access 9:106981–106997

    Article  Google Scholar 

  15. Beghdadi A, Mallem M (2022) A comprehensive overview of dynamic visual SLAM and deep learning: concepts, methods and challenges. Machine Vision and Applications 33(4):1–28

    Article  Google Scholar 

  16. Saputra MRU, Markham A, Trigoni N (2018) Visual SLAM and structure from motion in dynamic environments: A survey. ACM Computing Surveys (CSUR) 51(2):1–36

    Article  Google Scholar 

  17. Macario Barros A, Michel M, Moline Y, Corre G, Carrel F (2022) A comprehensive survey of visual slam algorithms. Robotics 11(1):24

    Article  Google Scholar 

  18. Cui L, Ma C (2020) SDF-SLAM: Semantic depth filter SLAM for dynamic environments. IEEE Access 8:95301–95311

    Article  Google Scholar 

  19. Zhang, T, Zhang, H, Li, Y, Nakamura, Y, & Zhang, L (2020) Flowfusion: Dynamic dense rgb-d slam based on optical flow. In 2020 IEEE international conference on robotics and automation (ICRA) (pp 7322–7328) IEEE

  20. Bolya, D, Zhou, C, Xiao, F, & Lee, YJ (2020) Yolact++: Better real-time instance segmentation. IEEE transactions on pattern analysis and machine intelligence

  21. Sturm, J, Engelhard, N, Endres, F, Burgard, W, & Cremers, D (2012) A benchmark for the evaluation of RGB-D SLAM systems. In 2012 IEEE/RSJ international conference on intelligent robots and systems (pp 573–580) IEEE

Download references

Acknowledgements

This work was supported by our IBISC’Lab and by the Strategic Research Initiatives project iCODE accredited by Paris Saclay University. REC-2019-011-AV-2 Strategic Research Initiatives - ICode Project Funding Agreement - Addendum 2 n°REC-2019-011-AV-2

Funding

This work was supported by our IBISC’Lab and by the Strategic Research Initiatives project iCODE accredited by Paris Saclay University. REC-2019-011-AV-2 Strategic Research Initiatives - ICode Project Funding Agreement - Addendum 2 n°REC-2019-011-AV-2

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Beghdadi Ayman.

Ethics declarations

Conflicts of interest

All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Malik Mallem and Lotfi Beji These authors contributed equally to this work.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ayman, B., Malik, M. & Lotfi, B. DAM-SLAM: depth attention module in a semantic visual SLAM based on objects interaction for dynamic environments. Appl Intell 53, 25802–25815 (2023). https://doi.org/10.1007/s10489-023-04720-3

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04720-3

Keywords

Navigation