Abstract
Nowadays, modern visual SLAM methods attempt to deal with dynamic environments by considering the non-rigid scene assumption. This well-established approach combines geometric and semantic information to detect dynamic objects to achieve accurate localization and mapping in real environments. However, these methods need more generalization and scene awareness because of their reasoning limits due to the labeling object strategy and the need for matched keypoints. Therefore, we propose a novel method called Depth Attention Module SLAM (DAM-SLAM) that overcomes the limitations of existing methods. The main idea is to take into account the depth influence used in the geometric and semantic modules through a depth-related adaptive threshold and impact factor. Moreover, a Bayesian filter is used to refine the keypoints state estimates using a motion probability update function based on a weighting strategy related to the keypoints area (in/out of segmented object’s masks). In addition, we designed a Depth Attention Module that allows generalization to other methods by considering the non-matched keypoints and the keypoints out of segmented regions. This module estimates these keypoints state without requiring any prior semantic information by determining the interactions between the objects. We estimate this interaction through the correlation between the proximity of depth and position of these keypoints with the dynamic keypoints in a specific zone of influence of dynamic objects. The obtained results demonstrate the efficacy of the proposed method in providing accurate localization and mapping in dynamic environments.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Davison AJ, Reid ID, Molton ND, Stasse O (2007) MonoSLAM: Real-time single camera SLAM. IEEE transactions on pattern analysis and machine intelligence 29(6):1052–1067
Klein, G, & Murray, D (2007) Parallel tracking and mapping for small AR workspaces. In 2007 6th IEEE and ACM international symposium on mixed and augmented reality (pp 225-234) IEEE
Mur-Artal R, Montiel JMM, Tardos JD (2015) ORB-SLAM: a versatile and accurate monocular SLAM system. IEEE transactions on robotics 31(5):1147–1163
Campos Carlos et al (2021) Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Transactions on Robotics 37(6):1874–1890
Mur-Artal R, Tardós JD (2017) Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE transactions on robotics 33(5):1255–1262
Qin T, Li P, Shen S (2018) Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Transactions on Robotics 34(4):1004–1020
Bescos B, Fácil JM, Civera J, Neira J (2018) DynaSLAM: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters 3(4):4076–4083
Xiao L, Wang J, Qiu X, Rong Z, Zou X (2019) Dynamic-SLAM: Semantic monocular visual localization and mapping based on deep learning in dynamic environment. Robotics and Autonomous Systems 117:1–16
Yu, C, Liu, Z, Liu, X. J, Xie, F, Yang, Y, Wei, Q, & Fei, Q (2018) DS-SLAM: A semantic visual SLAM towards dynamic environments. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp 1168–1174) IEEE
Yan L, Hu X, Zhao L, Chen Y, Wei P, Xie H (2022) DGS-SLAM: a fast and robust rgbd slam in dynamic environments combined by geometric and semantic information. Remote Sensing 14(3):795
Cheng J, Wang Z, Zhou H, Li L, Yao J (2020) DM-SLAM: A feature-based SLAM system for rigid dynamic scenes. ISPRS International Journal of Geo-Information 9(4):202
Cui L, Ma C (2019) SOF-SLAM: A semantic visual SLAM for dynamic environments. IEEE access 7:166528–166539
Li A, Wang J, Xu M, Chen Z (2021) DP-SLAM: A visual SLAM with moving probability towards dynamic environments. Information Sciences 556:128–142
Liu Y, Miura J (2021) RDMO-SLAM: Real-time visual SLAM for dynamic environments using semantic label prediction with optical flow. IEEE Access 9:106981–106997
Beghdadi A, Mallem M (2022) A comprehensive overview of dynamic visual SLAM and deep learning: concepts, methods and challenges. Machine Vision and Applications 33(4):1–28
Saputra MRU, Markham A, Trigoni N (2018) Visual SLAM and structure from motion in dynamic environments: A survey. ACM Computing Surveys (CSUR) 51(2):1–36
Macario Barros A, Michel M, Moline Y, Corre G, Carrel F (2022) A comprehensive survey of visual slam algorithms. Robotics 11(1):24
Cui L, Ma C (2020) SDF-SLAM: Semantic depth filter SLAM for dynamic environments. IEEE Access 8:95301–95311
Zhang, T, Zhang, H, Li, Y, Nakamura, Y, & Zhang, L (2020) Flowfusion: Dynamic dense rgb-d slam based on optical flow. In 2020 IEEE international conference on robotics and automation (ICRA) (pp 7322–7328) IEEE
Bolya, D, Zhou, C, Xiao, F, & Lee, YJ (2020) Yolact++: Better real-time instance segmentation. IEEE transactions on pattern analysis and machine intelligence
Sturm, J, Engelhard, N, Endres, F, Burgard, W, & Cremers, D (2012) A benchmark for the evaluation of RGB-D SLAM systems. In 2012 IEEE/RSJ international conference on intelligent robots and systems (pp 573–580) IEEE
Acknowledgements
This work was supported by our IBISC’Lab and by the Strategic Research Initiatives project iCODE accredited by Paris Saclay University. REC-2019-011-AV-2 Strategic Research Initiatives - ICode Project Funding Agreement - Addendum 2 n°REC-2019-011-AV-2
Funding
This work was supported by our IBISC’Lab and by the Strategic Research Initiatives project iCODE accredited by Paris Saclay University. REC-2019-011-AV-2 Strategic Research Initiatives - ICode Project Funding Agreement - Addendum 2 n°REC-2019-011-AV-2
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Malik Mallem and Lotfi Beji These authors contributed equally to this work.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Ayman, B., Malik, M. & Lotfi, B. DAM-SLAM: depth attention module in a semantic visual SLAM based on objects interaction for dynamic environments. Appl Intell 53, 25802–25815 (2023). https://doi.org/10.1007/s10489-023-04720-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-04720-3