ABSTRACT
Localization and mapping in the real scene where movable objects exist is a challenging task, however, most SLAM approaches generally assume that the scene is static. This assumption leads to building a map with movable objects, causing the localization to fail. We propose MODS, a movable object discrimination strategy which generates a knowledge base containing different attributes of an object, classify the object according to various attributes in the knowledge base, and take different measures for different categories. By using MODS, lidar data can be used in a selective manner to avoid data redundancy and reduce the probability of localization failure. We implement a complete SLAM system based on MODS, verify the influence of movable objects on SLAM with KITTI dataset, and also estimate our SLAM system. The qualitative experimental results show that our SLAM system can effectively eliminate the influence of movable objects on mapping and localization. Moreover, the quantitative experiment results show that the time required for localization was reduced by 16.0% on average compared with the state-of-the-art, while the loop-closure can still be correctly detected.
- Bosse, M. and Zlot, R. 2013. Place recognition using keypoint voting in large 3D lidar datasets. In 2013 IEEE International Conference on Robotics and Automation. 2677--2684.Google Scholar
- Bueno, M., Mart´inez-Sanchez, J., Gonz´alez-Jorge, H., and Lorenzo, H. 2016. Detection of Geometric Keypoints and its Application to Point Cloud Coarse Registration. ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences(2016), 187--194.Google Scholar
- Deschaud, J. 2018. IMLS-SLAM: Scan-to-Model Matching Based on 3D Data. In 2018 IEEE International Conference on Robotics and Automation (2018), 2480--2485.Google ScholarCross Ref
- Dewan, A., Caselitz, T., and Burgard, W. 2018. Learning a local feature descriptor for 3d lidar scans. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (2018), 4774--4780.Google Scholar
- Douillard, B., Underwood, J., Kuntz, N., Vlaskine, V., Quadros, A., Morton, P., and Frenkel, A. 2011. On the segmentation of 3d lidar point clouds. In 2011 IEEE International Conference on Robotics and Automation, (2011), 2798--2805.Google ScholarCross Ref
- Dub, R., Dugas, D., Stumm, E., Nieto, J., Siegwart, R., and Cadena, C. 2017. Segmatch: Segment based place recognition in 3d point clouds. In 2017 IEEE International Conference on Robotics and Automation (ICRA), 5266--5272.Google Scholar
- Fischler, M.A., and Bolles, R.C. 1981. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24, 6 (June 1981), 381--395. DOI=https://doi.org/10.1145/358669.358692 Google ScholarDigital Library
- Fuentes-Pacheco, J., Ascencio, J., and Rendon-Mancha, J. 2015. Visual simultaneous localization and mapping: A survey. Artificial Intelligence Review 43, 1 (01 January 2015), 55--81. DOI=http://dx.doi.org/10.1007/s10462-012-9365-8 Google ScholarDigital Library
- Geiger, A., Lenz, P., and Urtasun, R. 2012. Are we ready for autonomous driving? the kitti vision benchmark suite. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, 3354--3361. Google ScholarDigital Library
- Grisetti, G., Kummerle, R., Stachniss, C., and Burgard, W. 2010. A tutorial on graph-based slam. IEEE Intelligent Transportation Systems Magazine 2, 4 (2010), 31--43.Google ScholarCross Ref
- Hayman and Eklundh. 2003. Statistical background subtraction for a mobile observer. In Proceedings Ninth IEEE International Conference on Computer Vision, 67--74 vol.1. Google ScholarDigital Library
- Kim, G. and Kim, A. 2018. Scan context: Egocentric spatial descriptor for place recognition within 3d point cloud map. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 4802--4809.Google Scholar
- Nister, D. 2013. An efficient solution to the five-point relative pose problem. In 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings, Vol.2, II-195.Google Scholar
- Olson, C., H. Matthies, L., Schoppers, M., and W. Maimone, M. 2003. Rover navigation using stereo ego-motion. Robotics and Autonomous Systems 43, 4 (June 2003), 215--229.Google ScholarCross Ref
- Se, S., Lowe, D., and Little, J. 2002. Mobile robot localization and mapping with uncertainty using scale-invariant visual landmarks. The International Journal of Robotics Research 21, 8 (2002), 735--758.Google ScholarCross Ref
- Strasdat, H., Montiel, J.M.M., and Davison, A.J. 2010. Scale drift-aware large scale monocular slam. In Matsuoka, Y., Durrant-Whyte, H.F., Neira, J. (eds.) Robotics: Science and Systems. The MIT Press.Google Scholar
- Swadzba, A., Beuter, N., Wachsmuth, S., and Kummert, F. 2010. Dynamic 3d scene analysis for acquiring articulated scene models. In 2010 IEEE International Conference on Robotics and Automation, 134--141.Google Scholar
- Terashima, T. and Hasegawa, O. 2017. A visual-slam for first person vision and mobile robots. In 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), 73--76.Google Scholar
- Weinmann, M., Jutzi, B., and Mallet, C. 2013. Feature relevance assessment for the semantic interpretation of 3D point cloud data. ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences (2013), 313--318.Google Scholar
- Weinmann, M., Jutzi, B., and Mallet, C. 2014. Semantic 3D scene interpretation: A framework combining optimal neighborhood size selection with relevant features. ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences, (2014), 181--188. https://doi.org/10.5194/isprsannals-II-3-181-2014.Google Scholar
- Wu, B., Wan, A., Yue, X., and Keutzer, K. 2018. Squeezeseg: Convolutional neural nets with recurrent crf for real-time road-object segmentation from 3d lidar point cloud. In 2018 IEEE International Conference on Robotics and Automation (ICRA), 1887--1893. https://doi.org/10.1109/ICRA.2018.8462926Google Scholar
Index Terms
- StaticSLAM: A Dynamic Object-free Accurate SLAM System
Recommendations
SLAM-based localization of 3D gaze using a mobile eye tracker
ETRA '18: Proceedings of the 2018 ACM Symposium on Eye Tracking Research & ApplicationsPast work in eye tracking has focused on estimating gaze targets in two dimensions (2D), e.g. on a computer screen or scene camera image. Three-dimensional (3D) gaze estimates would be extremely useful when humans are mobile and interacting with the ...
SLAM and navigation in indoor environments
PSIVT'11: Proceedings of the 5th Pacific Rim conference on Advances in Image and Video Technology - Volume Part IIn this paper, we propose a system for wheeled robot SLAM and navigation in indoor environments. An omni-directional camera and a laser range finder are the sensors to extract the point features and the line features as the landmarks. In SLAM and self-...
ROS-Based SLAM for a Gazebo-Simulated Mobile Robot in Image-Based 3D Model of Indoor Environment
ACIVS 2015: Proceedings of the 16th International Conference on Advanced Concepts for Intelligent Vision Systems - Volume 9386Nowadays robot simulators have robust physics engines, high-quality graphics, and convenient interfaces, affording researchers to substitute physical systems with their simulation models in order to pre-estimate the performance of theoretical findings ...
Comments