Skip to main content

A Visually Impaired Assistant Using Neural Network and Image Recognition with Physical Navigation

  • Conference paper
  • First Online:
Advances in Neural Networks – ISNN 2020 (ISNN 2020)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 12557))

Included in the following conference series:

Abstract

In Hong Kong, over 2.4% of the total population suffered from visual impairment. They are facing many difficulties in their daily lives, such as shopping and travelling from places to places within the city. For outdoor activities, they usually need to have an assistant to guide their ways to reach the destinations. In this paper, a mobile application assisting visually impaired people for outdoor navigation is proposed. The application consists of navigation, obstacle detection and scene description functions. The navigation function assists the user to travel to the destination with the Global Positioning System (GPS) and sound guidance. The obstacle detection function alerts the visually impaired people for any obstacles ahead that may be avoided for collision. The scene description function describes the scene in front of the users with voice. In general, the mobile application can assist the people with low vision to walk on the streets safely, reliably and efficiently.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://eye-d.in/.

  2. 2.

    https://play.google.com/store/apps/details?id=hk.com.redso.read4u&hl=zh_HK.

  3. 3.

    https://www.facebook.com/InnoTechAssociation/.

  4. 4.

    https://play.google.com/store/apps/details?id=com.hkblindunion.smartcitywalk.android&hl=zh_HK.

  5. 5.

    https://www.hkbu.org.hk/.

  6. 6.

    https://wewalk.io/en/.

  7. 7.

    http://mqtt.org/.

  8. 8.

    https://flask.palletsprojects.com/en/1.1.x/.

  9. 9.

    https://cloud.google.com/speech-to-text.

  10. 10.

    https://www.mapbox.com/.

References

  1. Vos, T., et al.: Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: a systematic analysis for the Global Burden of Disease Study 2015. Lancet 388(10053), 1545–1602 (2016)

    Article  Google Scholar 

  2. Statistics on People with Visual Impairment. https://www.hkbu.org.hk/en/knowledge/statistics/index

  3. Image Description-Computer Vision - Azure Cognitive Services. https://docs.microsoft.com/zh-tw/azure/cognitive-services/computer-vision/concept-describing-images

  4. Pricing-Computer Vision API. https://azure.microsoft.com/en-gb/pricing/details/cognitive-services/computer-vision/

  5. Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: International Conference on Machine Learning, pp. 2048–2057 (2015)

    Google Scholar 

  6. Vinyals, O., Toshev, A., Bengio, S., Erhan, D.: Show and tell: a neural imagecaption generator. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2015). https://doi.org/10.1109/cvpr.2015.7298935

  7. Graves, A., Navdeep, J.: Towards end-to-end speech recognition with recurrent neural networks. In: International Conference on Machine Learning, pp. 1764–1772 (2014)

    Google Scholar 

  8. He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961–2969 (2017)

    Google Scholar 

  9. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: MobileNetV2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510–4520 (2018)

    Google Scholar 

  10. Alhashim, I., Wonka, P.: High quality monocular depth estimation via transfer learning. arXiv preprint arXiv:1812.11941 (2018)

  11. Newman, N.: Apple iBeacon technology briefing. J. Direct Data Digit. Mark. Pract. 15(3), 222–225 (2014). https://doi.org/10.1057/dddmp.2014.7

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Sin-Chun Ng or Chok-Pang Kwok .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, OC.A., Li, SK., Yan, LQ., Ng, SC., Kwok, CP. (2020). A Visually Impaired Assistant Using Neural Network and Image Recognition with Physical Navigation. In: Han, M., Qin, S., Zhang, N. (eds) Advances in Neural Networks – ISNN 2020. ISNN 2020. Lecture Notes in Computer Science(), vol 12557. Springer, Cham. https://doi.org/10.1007/978-3-030-64221-1_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-64221-1_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-64220-4

  • Online ISBN: 978-3-030-64221-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics