Skip to main content
Log in

DenseFilter: Feature Correspondence Filter Based on Dense Networks for VSLAM

  • Regular paper
  • Published:
Journal of Intelligent & Robotic Systems Aims and scope Submit manuscript

Abstract

Feature matching can establish reliable points correspondences between two images, which plays a vital role in indirect Visual Simultaneous Localization and Mapping (VSLAM). The VSLAM system performs data association between image frames through feature matching. There are a large number of incorrect correspondences in the unfiltered matching results, and the accuracy of data association will directly determine the localization performance. This paper proposes a convolutional neural network with dense connectivity to eliminate the outliers in the matching results. It repeatedly concatenates low-level and high-level features into a dense feature map, which allows both local and global information to be used for mismatch removal. Furthermore, we introduce an outlier rejection module based on our network into the classic monocular visual odometry (VO) pipeline, in order to enhance the reliability of data association. Experiments on several benchmark datasets show that our method achieves favorable performance against other algorithms, especially under challenging scenes, such as low texture. Also, we verify the performance of our system on a mobile robot in real-world localization task.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Data Availability

Not applicable.

Code Availability

Not applicable.

References

  1. Shi, J.: Good features to track. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp. 593–600 (1994)

  2. Rosten, E., Drummond, T.: Machine learning for high-speed corner detection. In: European Conference on Computer Vision, pp. 430–443 (2006)

  3. Rublee, E., Rabaud, V., Konolige, K., Bradski, G.: ORB: An efficient alternative to SIFT or SURF. In: 2011 International Conference on Computer Vision, pp. 2564–2571 (2011)

  4. Davison, A.J., Reid, I.D., Molton, N.D., Stasse, O.: MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Machine Intell. 29(6), 1052–1067 (2007)

    Article  Google Scholar 

  5. Klein, G., Murray, D.: Parallel tracking and mapping for small AR workspaces. In: 2007 6th IEEE and ACM international symposium on mixed and augmented reality, pp. 225–234 (2007)

  6. Mur-Artal, R., Tardós, J.D.: ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 33(5), 1255–1262 (2017)

    Article  Google Scholar 

  7. Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60(2), 91–110 (2004)

    Article  Google Scholar 

  8. Collins, T., Mesejo, P., Bartoli, A.: An analysis of errors in graph-based keypoint matching and proposed solutions. In: European conference on computer vision, pp. 138–153 (2014)

  9. Lin, W.Y.D., Cheng, M.M., Lu, J., Yang, H., Do, M.N., Torr, P.: Bilateral functions for global motion modeling. In: European Conference on Computer Vision, pp. 341–356 (2014)

  10. Lin, W.Y., Wang, F., Cheng, M.M., Yeung, S.K., Torr, P.H., Do, M.N., Lu, J.: CODE: Coherence based decision boundaries for feature correspondence. IEEE Trans. Pattern Anal. Machine Intell. 40(1), 34–47 (2017)

    Article  Google Scholar 

  11. Bian, J., Lin, W.Y., Matsushita, Y., Yeung, S.K., Nguyen, T.D., Cheng, M.M.: Gms: Grid-based motion statistics for fast, ultra-robust feature correspondence. In: Proc. IEEE conference on computer vision and pattern recognition, pp. 4181–4190 (2017)

  12. Jiang, X., Ma, J., Jiang, J., Guo, X.: Robust feature matching using spatial clustering with heavy outliers. IEEE Trans. Image Process. 29, 736–746 (2019)

    Article  MathSciNet  Google Scholar 

  13. Ma, J., Zhao, J., Jiang, J., Zhou, H., Guo, X.: Locality preserving matching. Int. J. Comput. Vis. 127(5), 512–531 (2019)

    Article  MathSciNet  Google Scholar 

  14. Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381–395 (1981)

    Article  MathSciNet  Google Scholar 

  15. Torr, P.H., Zisserman, A.: MLESAC: A new robust estimator with application to estimating image geometry. Comput. Vis. Image Underst. 78(1), 138–156 (2000)

    Article  Google Scholar 

  16. Chum, O., Matas, T.: Matching with PROSAC - progressive sample consensus. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), vol.1, pp. 220–226 (2005)

  17. Raguram, R., Chum, O., Pollefeys, M., Matas, J., Frahm, J.M.: USAC: A universal framework for random sample consensus. IEEE Trans. Pattern Anal. Machine Intell. 35(8), 2022–2038 (2012)

    Article  Google Scholar 

  18. Brachmann, E., Rother, C.: Neural-guided RANSAC: Learning where to sample model hypotheses. In: Proc. IEEE/CVF International Conference on Computer Vision, pp. 4322–4331 (2019)

  19. Lin, T., Wang, X.: Hierarchical clustering matching for features with repetitive patterns in visual odometry. J. Intell. Robot Syst. 100(3), 1139–1155 (2020)

    Article  Google Scholar 

  20. Zhao, Y., Vela, P.A.: Good feature selection for least squares pose optimization in VO/VSLAM. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1183–1189 (2018)

  21. Yi, K.M., Trulls, E., Ono, Y., Lepetit, V., Salzmann, M., Fua, P.: Learning to find good correspondences. In: Proc. IEEE conference on computer vision and pattern recognition, pp. 2666–2674 (2018)

  22. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proc. IEEE conference on computer vision and pattern recognition, pp. 770–778 (2016)

  23. Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proc. IEEE conference on computer vision and pattern recognition, pp. 4700–4708 (2017)

  24. Longuet-Higgins, H.C.: A computer algorithm for reconstructing a scene from two projections. Nature 293(5828), 133–135 (1981)

    Article  Google Scholar 

  25. Thomee, B., Shamma, D. A., Friedland, G., Elizalde, B., Ni, K., Poland, D., … Li,L. J.: YFCC100M: The new data in multimedia research. Commun. ACM. 59(2), 64–73 (2016)

  26. Xiao, J., Owens, A., Torralba, A.: Sun3d: A database of big spaces reconstructed using sfm and object labels. In Proc. IEEE international conference on computer vision, pp. 1625–1632 (2013)

  27. Sturm, J., Engelhard, N., Endres, F., Burgard, W., Cremers, D.: A benchmark for the evaluation of RGB-D SLAM systems. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 573–580 (2012)

  28. Burri, M., Nikolic, J., Gohl, P., Schneider, T., Rehder, J., Omari, S., … Siegwart, R.: The EuRoC micro aerial vehicle datasets. Int. J. Rob. Res. 35(10), 1157–1163 (2016)

  29. Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: The kitti dataset. Int. J. Rob. Res. 32(11), 1231–1237 (2013)

    Article  Google Scholar 

Download references

Funding

This work was supported by Department of science and technology of Guangdong Province (No:2021B01420003) and Department of science and technology of Foshan Nanhai District (No: 201811020005).

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the methodology conception and design. Data collection and analysis were performed by Jinze Xu and Lingfeng Su. Jinze Xu, Lingfeng Su and Feng Ye wrote the manuscript. Kuo Li and Yizong Lai reviewed and approved the manuscript.

Corresponding author

Correspondence to Feng Ye.

Ethics declarations

Ethical Approval

Not applicable.

Consent to Participate

Not applicable.

Consent to Publish

Not applicable.

Conflict of Interest

The authors declare that they do not have any conflicts of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xu, J., Su, L., Ye, F. et al. DenseFilter: Feature Correspondence Filter Based on Dense Networks for VSLAM. J Intell Robot Syst 106, 18 (2022). https://doi.org/10.1007/s10846-022-01735-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10846-022-01735-9

Keywords

Navigation