Skip to main content

Stereo Based 3D Perception for Obstacle Avoidance in Autonomous Wheelchair Navigation

  • Conference paper
  • First Online:
ROBOT2022: Fifth Iberian Robotics Conference (ROBOT 2022)

Abstract

In recent years, scientific and technological advances in robotics, have enabled the development of disruptive solutions for human interaction with the real world. In particular, the application of robotics to support people with physical disabilities, improved their life quality with a high social impact. This paper presents a stereo image based perception solution for autonomous wheelchairs navigation. It was developed to extend the Intellwheels project, a development platform for intelligent wheelchairs. The current version of Intellwheels relies on a planar scanning sensor, the Laser Range Finder (LRF), to detect the surrounding obstacles. The need for robust navigation capabilities means that the robot is required to precept not only obstacles but also bumps and holes on the ground. The proposed stereo-based solution, supported in passive stereo ZED cameras, was evaluated in a 3D simulated world scenario designed with a challenging floor. The performance of the wheelchair navigation with three different configurations was compared: first, using a LRF sensor, next with an unfiltered stereo camera and finally, applying a stereo camera with a speckle filter. The LRF solution was unable to complete the planned navigation. The unfiltered stereo camera completed the challenge with a low navigation quality due to noise. The filtered stereo camera reached the target position with a nearly optimal path.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Fauadi, M.H.F.M., Ali, M., Ramlan, S., Noor, A.Z.M.: Intelligent vision-based navigation system for mobile robot: a technological review. Periodicals Eng. Nat. Sci. (PEN) 6, 47–57 (2018). https://doi.org/10.21533/pen.v6i2.174

    Article  Google Scholar 

  2. LIACC, Intellwheels (2012). https://liacc.fe.up.pt/project/intellwheels. Accessed 24 Jan 2022

  3. Campbell, S., et al.: Sensor technology in autonomous vehicles: a review. In: 2018 29th Irish Signals and Systems Conference (ISSC), pp. 1–4 (2018). https://doi.org/10.1109/ISSC.2018.8585340

  4. Vargas, J., Alsweiss, S., Toker, O., Razdan, R., Santos, J.: An overview of autonomous vehicles sensors and their vulnerability to weather conditions. Sensors (Basel, Switzerland), vol. 21. https://doi.org/10.3390/s21165397

  5. Alenyà, G., Foix, S., Torras, C.: Using ToF and RGBD cameras for 3D robot perception and manipulation in human environments. Intell. Serv. Robot. 7(4), 211–220 (2014). https://doi.org/10.1007/s11370-014-0159-5

    Article  Google Scholar 

  6. Bianco, G., Gallo, A., Bruno, F., Muzzupappa, M.: A comparison between active and passive techniques for underwater 3d applications. In: ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. XXXVIII-5/W16, pp. 357–363 (2012). https://doi.org/10.5194/isprsarchives-XXXVIII-5-W16-357-2011

  7. Nardi, F., Lázaro, M.T., Iocchi, L., Grisetti, G.: Generation of laser-quality 2D navigation maps from RGB-D sensors. In: Holz, D., Genter, K., Saad, M., von Stryk, O. (eds.) RoboCup 2018. LNCS (LNAI), vol. 11374, pp. 238–250. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-27544-0_20

    Chapter  Google Scholar 

  8. Widodo, N.S., Pamungkas, A.: Machine vision-based obstacle avoidance for mobile robot. J. Ilm. Tek. Elektro Komput. dan Inform. 5(2), 77–84 (2020). https://doi.org/10.26555/jiteki.v5i2.14767

    Article  Google Scholar 

  9. Nagarajan, V.R., Singh, P.: Obstacle detection and avoidance for mobile robots using monocular vision. In: 2021 8th International Conference on Smart Computing and Communications (ICSCC), pp. 275–279 (2021). https://doi.org/10.1109/ICSCC51209.2021.9528162

  10. Son, T.J., Hassan, A.H.A., Jairan, M.H.: Optimized robot mapping and obstacle avoidance using stereo vision. In: 2021 IEEE 11th IEEE Symposium on Computer Applications Industrial Electronics (ISCAIE), pp. 279–284 (2021). https://doi.org/10.1109/ISCAIE51753.2021.9431769

  11. Park, K.: Sideguide: a large-scale sidewalk dataset for guiding impaired people. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 10022–10029 (2020). https://doi.org/10.1109/IROS45743.2020.9340734

  12. Koenig, N., Howard, A.: Design and use paradigms for gazebo, an open-source multi-robot simulator. In: 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)(IEEE Cat. No. 04CH37566), vol. 3, pp. 2149–2154. IEEE (2004)

    Google Scholar 

  13. ROS.org, Ros navigation stack (2020). http://wiki.ros.org/navigation. Accessed 24 Jan 2022

  14. Held, R., Banks, M.: Misperceptions in stereoscopic displays: a vision science perspective. In: ACM Transactions on Graphics 2008, pp. 23–32 (2008). https://doi.org/10.1145/1394281.1394285

  15. ROS.org, Ros unified robot description format (urdf) (2019). http://wiki.ros.org/urdf. Accessed 24 Jan 2022

  16. Stereolabs, Zed, 2 (2021). https://www.stereolabs.com/zed-2/. Accessed 24 Jan 2022

  17. Macenski, S., Tsai, D., Feinberg, M.: Spatio-temporal voxel layer: a view on robot perception for the dynamic world. Int. J. Adv. Robot. Syst. 17, 172988142091053 (2020). https://doi.org/10.1177/1729881420910530

  18. ROS.org, Choosing good stereo parameters (2018). http://wiki.ros.org/stereo_image_proc/Tutorials/ChoosingGoodStereoParameters. Accessed 09 July 2022

  19. ROS.org, Setup and configuration of the navigation stack on a robot (2018). http://wiki.ros.org/navigation/Tutorials/RobotSetup. Accessed 24 Jan 2022

  20. Zeineldin, R., El-Fishawy, N.: Fast and accurate ground plane detection for the visually impaired from 3d organized point clouds. In: 2016 SAI Computing Conference (SAI), pp. 373–379 (2016). https://doi.org/10.1109/SAI.2016.7556009

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bruno Gomes .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Gomes, B., Torres, J., Sobral, P., Sousa, A., Reis, L.P. (2023). Stereo Based 3D Perception for Obstacle Avoidance in Autonomous Wheelchair Navigation. In: Tardioli, D., Matellán, V., Heredia, G., Silva, M.F., Marques, L. (eds) ROBOT2022: Fifth Iberian Robotics Conference. ROBOT 2022. Lecture Notes in Networks and Systems, vol 589. Springer, Cham. https://doi.org/10.1007/978-3-031-21065-5_27

Download citation

Publish with us

Policies and ethics