Skip to main content
Log in

User-centered system design for assisted navigation of visually impaired individuals in outdoor cultural environments

  • Long Paper
  • Published:
Universal Access in the Information Society Aims and scope Submit manuscript

Abstract

The marginalization of people with disabilities, such as visually impaired individuals (VIIs), has driven scientists to take advantage of the fast growth of smart technologies and develop smart assistive systems (SASs) to bring VIIs back to social life, education and even to culture. Our research focuses on developing a human–computer interactive system that will guide VIIs in outdoor cultural environments by offering universal access to cultural information, social networking and safe navigation among other services. The VI users interact with computer-based SAS to control the system during its operation, while having access to remote connection with non-VIIs for external guidance and company. The development of such a system needs a user-centered design (UCD) that incorporates the elicitation of the necessary requirements for a satisfying operation for the VI users. In this paper, we present a novel SAS system for VIIs and its design considerations, which follow a UCD approach to determine a set of operational, functional, ergonomic, environmental and optional requirements of the system. Both VIIs and non-VIIs took part in a series of interviews and questionnaires, from which data were analyzed to form the requirements of the system for both the on-site and remote use. The final requirements are tested by trials and their evaluation and results are presented. The experimental investigations gave significant feedback for the development of the system, throughout the design process. The most important contribution of this study is the derivation of requirements applicable not only to the specific system under investigation, but also to other relevant SASs for VIIs.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16

Similar content being viewed by others

References

  1. WHO: World Health Organization—blindness and visual impairement. http://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairment (2018). Accessed Sept 2019

  2. Vuletić, G., Šarlija, T., Benjak, T.: Quality of life in blind and partially sighted people. J. Appl. Health Sci. 2, 101–112 (2016)

    Google Scholar 

  3. Wahab, M.H.A., Talib, A.A., Kadir, H.A., Johari, A., Noraziah, A., Sidek, R.M., Mutalib, A.A.: Smart cane: assistive cane for visually-impaired people (2011). arXiv preprint arXiv:1110.5156

  4. Ulrich, I., Borenstein, J.: The GuideCane-applying mobile robot technologies to assist the visually impaired. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 31(2), 131–136 (2001)

    Google Scholar 

  5. Takizawa, H., Yamaguchi, S., Aoyagi, M., Ezaki, N., Mizuno, S.: Kinect cane: an assistive system for the visually impaired based on the concept of object recognition aid. Pers. Ubiquit. Comput. 19(5–6), 955–965 (2015)

    Google Scholar 

  6. Mandru, D., Lungu, I.B., Mociran, A.: Development of a mechatronic blind stick, fascicle of management and technological. Engineering 6(16), 797–802 (2007)

    Google Scholar 

  7. Gayathri, G., Vishnupriya, M., Nandhini, R., Banupriya, M.: Smart walking stick for visually impaired. IJECS 3(3), 4057–4061 (2014)

    Google Scholar 

  8. Nada, A.A., Fakhr, M.A., Seddik, A.F.: Assistive infrared sensor based smart stick for blind people. In: 2015 Science and Information Conference (SAI), pp. 1149–1154 (2015)

  9. Sakhardande, J., Pattanayak, P., Bhowmick, M.: Smart cane assisted mobility for the visually impaired. World Academy of Science, Engineering and Technology, vol. 70 (2012)

  10. “WeWALK”. https://wewalk.io/en (2019)

  11. Alghamdi, S., Van Schyndel, R., Alahmadi, A.: Indoor navigational aid using active RFID and QR-code for sighted and blind people. In: 2013 IEEE Eighth International Conference on Intelligent Sensors, Sensor Networks and Information Processing, pp. 18–22 (2013)

  12. Nassih, M., Cherradi, I., Maghous, Y., Ouriaghli, B., Salih-Alj, Y.: Obstacles recognition system for the blind people using RFID. In: 2012 Sixth International Conference on Next Generation Mobile Applications, Services and Technologies, pp. 60–63 (2012)

  13. Ivanov, R.: RSNAVI: an RFID-based context-aware indoor navigation system for the blind. In: Proceedings of the 13th International Conference on Computer Systems and Technologies, pp. 313–320 (2012)

  14. Qinghui, T., Malik, M.Y., Hong, Y., Park, J.: A real-time localization system using RFID for visually impaired (2011). arXiv preprint arXiv:1109.1879

  15. Yelamarthi, K., Haas, D., Nielsen, D., Mothersell, S.: RFID and GPS integrated navigation system for the visually impaired. In: 2010 53rd IEEE International Midwest Symposium on Circuits and Systems, pp. 1149–1152 (2010)

  16. Jafri, R., Ali, S.A., Arabnia, H.R.: Computer vision-based object recognition for the visually impaired using visual tags. In: Proceedings of the International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV), p. 1 (2013)

  17. Hakobyan, L., Lumsden, J., O’Sullivan, D., Bartlett, H.: Mobile assistive technologies for the visually impaired. Surv. Ophthalmol. 58(6), 513–528 (2013)

    Google Scholar 

  18. Liu, K.-C., Wu, C.-H., Tseng, S.-Y., Tsai, Y.-T.: Voice helper: a mobile assistive system for visually impaired persons. In: 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, pp. 1400–1405 (2015)

  19. Mekhalfi, M.L., Melgani, F., Zeggada, A., De Natale, F.G., Salem, M.A.-M., Khamis, A.: Recovering the sight to blind people in indoor environments with smart technologies. Expert Syst. Appl. 46, 129–138 (2016)

    Google Scholar 

  20. Tian, Y., Yang, X., Arditi, A.: Computer vision-based door detection for accessibility of unfamiliar environments to blind persons. In: International Conference on Computers for Handicapped Persons, pp. 263–270 (2010)

  21. Tian, Y., Yang, X., Yi, C., Arditi, A.: Toward a computer vision-based wayfinding aid for blind persons to access unfamiliar indoor environments. Mach. Vis. Appl. 24(3), 521–535 (2013)

    Google Scholar 

  22. Tee, Z., Ang, L., Seng, K.: Smart guide system to assist visually impaired people in an indoor environment. Iete Tech. Rev. 27(6), 455–464 (2010)

    Google Scholar 

  23. Jafri, R., Ali, S.A., Arabnia, H.R., Fatima, S.: Computer vision-based object recognition for the visually impaired in an indoors environment: a survey. Visual Comput. 30(11), 1197–1222 (2014)

    Google Scholar 

  24. Bai, J., Lian, S., Liu, Z., Wang, K., Liu, D.: Smart guiding glasses for visually impaired people in indoor environment. IEEE Trans. Consum. Electron. 63(3), 258–266 (2017)

    Google Scholar 

  25. He, H., Li, Y., Guan, Y., Tan, J.: Wearable ego-motion tracking for blind navigation in indoor environments. IEEE Trans. Autom. Sci. Eng. 12(4), 1181–1190 (2015)

    Google Scholar 

  26. Owayjan, M., Hayek, A., Nassrallah, H., Eldor, M.: Smart assistive navigation system for blind and visually impaired individuals. In: 2015 International Conference on Advances in Biomedical Engineering (ICABME), pp. 162–165 (2015)

  27. Wang, H.-C., Katzschmann, R.K., Teng, S., Araki, B., Giarré, L., Rus, D.: Enabling independent navigation for visually impaired people through a wearable vision-based feedback system. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 6533–6540 (2017)

  28. Mascetti, S., Ahmetovic, D., Gerino, A., Bernareggi, C.: ZebraRecognizer: pedestrian crossing recognition for people with visual impairment or blindness. Pattern Recognit. 60, 405–419 (2016)

    Google Scholar 

  29. Wang, S., Pan, H., Zhang, C., Tian, Y.: RGB-D image-based detection of stairs, pedestrian crosswalks and traffic signs. J. Vis. Commun. Image Represent. 25(2), 263–272 (2014)

    Google Scholar 

  30. Wang, S., Tian, Y.: Detecting stairs and pedestrian crosswalks for the blind by RGBD camera. In: 2012 IEEE International Conference on Bioinformatics and Biomedicine Workshops, pp. 732–739 (2012)

  31. Jafri, R., Ali, S.A.: Exploring the potential of eyewear-based wearable display devices for use by the visually impaired. In: 2014 3rd International Conference on User Science and Engineering (i-USEr), pp. 119–124 (2014)

  32. Shinohara, K., Wobbrock, J.O.: In the shadow of misperception: assistive technology use and social interactions. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 705–714 (2011)

  33. Law, C.M., Jaeger, P.T., McKay, E.: User-centered design in universal design resources? Univ. Access Inf. Soc. 9(4), 327–335 (2010)

    Google Scholar 

  34. Mirri, S., Roccetti, M., Salomoni, P.: Collaborative design of software applications: the role of users. Hum. Centric Comput. Inf. Sci. 8(1), 6 (2018)

    Google Scholar 

  35. Lopes, A., Valentim, N., Moraes, B., Zilse, R., Conte, T.: Applying user-centered techniques to analyze and design a mobile application. J. Softw. Eng. Res. Dev. 6(1), 5 (2018)

    Google Scholar 

  36. Vesin, B., Mangaroska, K., Giannakos, M.: Learning in smart environments: user-centered design and analytics of an adaptive learning system. Smart Learn. Environ. 5(1), 24 (2018)

    Google Scholar 

  37. Seyff, N., Todoran, I., Caluser, K., Singer, L., Glinz, M.: Using popular social network sites to support requirements elicitation, prioritization and negotiation. J. Internet Serv. Appl. 6(1), 7 (2015)

    Google Scholar 

  38. Lin, Y.-C., Yeh, C.-H., Wei, C.-C.: How will the use of graphics affect visual aesthetics? A user-centered approach for web page design. Int. J. Hum. Comput. Stud. 71(3), 217–227 (2013)

    Google Scholar 

  39. Costa, N.A., Holder, E., MacKinnon, S.: Implementing human centred design in the context of a graphical user interface redesign for ship manoeuvring. Int. J. Hum. Comput. Stud. 100, 55–65 (2017)

    Google Scholar 

  40. Augusto, J., Kramer, D., Alegre, U., Covaci, A., Santokhee, A.: The user-centred intelligent environments development process as a guide to co-create smart technology for people with special needs. Univ. Access Inf. Soc. 17(1), 115–130 (2018)

    Google Scholar 

  41. Lazar, J.: Web Usability: A User-Centered Design Approach. Pearson Addison Wesley, Boston (2006)

    Google Scholar 

  42. Juárez-Ramírez, R.: User-Centered Design and Adaptive Systems: Toward Improving Usability and accessibility. Springer, Berlin (2017)

    Google Scholar 

  43. Jafri, R., Khan, M.M.: User-centered design of a depth data based obstacle detection and avoidance system for the visually impaired. Hum. Centric Comput. Inf. Sci. 8(1), 14 (2018)

    Google Scholar 

  44. Mancas-Thillou, C., Ferreira, S., Demeyer, J., Minetti, C., Gosselin, B.: A multifunctional reading assistant for the visually impaired. J. Image Video Process. 2007(3), 5 (2007)

    Google Scholar 

  45. Bateman, A., Zhao, O.K., Bajcsy, A.V., Jennings, M.C., Toth, B.N., Cohen, A.J., Horton, E.L., Khattar, A., Kuo, R.S., Lee, F.A., et al.: A user-centered design and analysis of an electrostatic haptic touchscreen system for students with visual impairments. Int. J. Hum. Comput. Stud. 109, 102–111 (2018)

    Google Scholar 

  46. Meza-de-Luna, M.E., Terven, J.R., Raducanu, B., Salas, J.: A Social-Aware Assistant to support individuals with visual impairments during social interaction: a systematic requirements analysis. Int. J. Hum. Comput. Stud. 122, 50–60 (2019)

    Google Scholar 

  47. Koo, S.H., Fallon, K.: Explorations of wearable technology for tracking self and others. Fashion Text. 5(1), 8 (2018)

    Google Scholar 

  48. Robins, B., Ferrari, E., Dautenhahn, K., Kronreif, G., Prazak-Aram, B., Gelderblom, G., Tanja, B., Caprino, F., Laudanna, E., Marti, P.: Human-centred design methods: developing scenarios for robot assisted play informed by user panels and field trials. Int. J. Hum. Comput. Stud. 68(12), 873–898 (2010)

    Google Scholar 

  49. Iakovidis, D.K., Diamantis, D., Dimas, D., Ntakolia, C., Spyrou, E.: Digital enhancement of cultural experience and accessibility for the visually impaired. In: Technological Trends in Improved Mobility of the Visually Impaired, pp. 237–271. Springer (2020)

  50. Loukopoulos, T., Koziri, M., Panagou, N., Papadopoulos, P.K., Iakovidis, D.K.: Cloud video guidance as ‘Deus ex Machina’ for the visually impaired. In: Technological Trends in Improved Mobility of the Visually Impaired. Springer, pp. 127–143 (2020)

  51. Zowghi, D., Coulin, C.: Requirements elicitation: a survey of techniques, approaches, and tools. In: Engineering and Managing Software Requirements, pp. 19–46. Springer (2005)

  52. Magnusson, C., Hedvall, P.-O., Caltenco, H.: Co-designing together with persons with visual impairments. In: Mobility of Visually Impaired People, pp. 411–434. Springer (2018)

  53. I. O. for Standardization, “ISO 9241-210:2010” (2010)

  54. Panchanathan, S., Black, J., Rush, M., Iyer, V.: iCare-a user centric approach to the development of assistive devices for the blind and visually impaired. In: 15th IEEE International Conference on Tools with Artificial Intelligence, 2003. Proceedings, pp. 641–648

  55. Fryer, L.: Putting it into words: the impact of visual impairment on perception, experience and presence. Goldsmiths, University of London (2013)

  56. Sosa-Garcia, J., Odone, F.: Hands on “Visual recognition for visually impaired users”. ACM Trans. Access. Comput. (TACCESS) 10(3), 8 (2017)

    Google Scholar 

  57. Panchanathan, S., Black, J., Rush, M., Iyer, V.: iCare-a user centric approach to the development of assistive devices for the blind and visually impaired. In: 15th IEEE International Conference on Tools with Artificial Intelligence, 2003. Proceedings, pp. 641–648 (2003)

  58. Perakovi, D., Periša, M., Prci, A.B.: Possibilities of applying ICT to improve safe movement of blind and visually impaired persons. In: Cutting Edge Research in Technologies. InTech (2015)

  59. Hersh, M.A., Johnson, M.A.: A robotic guide for blind people. Part 1. A multi-national survey of the attitudes, requirements and preferences of potential end-users. Appl. Bion. Biomech. 7(4), 277–288 (2010)

    Google Scholar 

  60. Zeng, L.: A survey: outdoor mobility experiences by the visually impaired. In: Mensch und Computer 2015—Workshopband (2015)

  61. Conradie, P., de Goedelaan, G.K., Mioch, T., Saldien, J.: Blind user requirements to support tactile mobility. In: Tactile Haptic User Interfaces for Tabletops and Tablets (TacTT 2014), pp. 48–53 (2014)

  62. Loomis, J.M., Marston, J.R., Golledge, R.G., Klatzky, R.L.: Personal guidance system for people with visual impairment: a comparison of spatial displays for route guidance. J. Visual Impair. Blindness 99(4), 219 (2005)

    Google Scholar 

  63. Petrie, H., Johnson, V., Strothotte, T., Raab, A., Fritz, S., Michel, R.: MOBIC: designing a travel aid for blind and elderly people. J. Navig. 49(01), 45–52 (1996)

    Google Scholar 

  64. Kulyukin, V., Gharpure, C., Nicholson, J., Pavithran, S.: RFID in robot-assisted indoor navigation for the visually impaired. In: 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No. 04CH37566) (2004)

  65. Chebat, D.-R., Schneider, F., Kupers, R., Ptito, M.: Navigation with a sensory substitution device in congenitally blind individuals. NeuroReport 22, 342–347 (2011)

    Google Scholar 

  66. Johnson, L.A., Higgins, C.M.: A navigation aid for the blind using tactile-visual sensory substitution. In: 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 6289–6292 (2006)

  67. Benabid, A., AlZuhair, M.: User involvement in the development of indoor navigation system for the visually impaired: a needs-finding study. In: 2014 3rd International Conference on User Science and Engineering (i-USEr), pp. 97–102 (2014)

  68. Alkhanifer, A., Ludi, S.: Towards a situation awareness design to improve visually impaired orientation in unfamiliar buildings: requirements elicitation study. In: 2014 IEEE 22nd International on Requirements Engineering Conference (RE), pp. 23–32 (2014)

  69. Kalia, A.A., Legge, G.E., Roy, R., Ogale, A.: Assessment of indoor route-finding technology for people with visual impairment. J. Visual Impair. Blindness 104(3), 135 (2010)

    Google Scholar 

  70. Roentgen, U.R., Gelderblom, G.J., de Witte, L.P.: User evaluation of two electronic mobility aids for persons who are visually impaired: a quasi-experimental study using a standardized mobility course. Assist. Technol. 24(2), 110–120 (2012)

    Google Scholar 

  71. Hub, A., Diepstraten, J., Ertl, T.: Design and development of an indoor navigation and object identification system for the blind. In: Proceedings of the ACM SIGACCESS Conference on Computers and Accessibility—ASSETS textquotesingle04 (2004)

  72. Guerrero, L.A., Vasquez, F., Ochoa, S.F.: An indoor navigation system for the visually impaired. Sensors 12(6), 8236–8258 (2012)

    Google Scholar 

  73. Strumillo, P.: Electronic interfaces aiding the visually impaired in environmental access, mobility and navigation. In: 3rd International Conference on Human System Interaction (2010)

  74. Bohonos, S., Lee, A., Malik, A., Thai, C., Manduchi, R.: Universal real-time navigational assistance (URNA). In: Proceedings of the 1st ACM SIGMOBILE International Workshop on Systems and Networking Support for Healthcare and Assisted Living Environments—HealthNet textquotesingle07 (2007)

  75. Sánchez, J., Maureira, E.: Subway mobility assistance tools for blind users. In: Universal Access in Ambient Intelligence, Lecture Notes on Computer Science, vol. 4397, pp. 386–404 (2007)

  76. Kim, H.K., Han, S.H., Park, J., Park, J.: The interaction experiences of visually impaired people with assistive technology: a case study of smartphones. Int. J. Ind. Ergon. 55, 22–33 (2016)

    Google Scholar 

  77. Kane, S.K., Jayant, C., Wobbrock, J.O., Ladner, R.E.: Freedom to roam: a study of mobile device adoption and accessibility for people with visual and motor disabilities. In: Proceedings of the 11th International ACM SIGACCESS Conference on Computers and Accessibility, pp. 115–122 (2009)

  78. Ungar, S., Blades, M., Spencer, C.: The construction of cognitive maps by children with visual impairments. In: Portugali, J. (ed.) The construction of cognitive maps, pp. 247–273. Springer Netherlands, Dordrecht (1996)

    Google Scholar 

  79. Periša, M., Cviti, I., Sente, R.E.: Social network customer requirements analysis for visually impaired people. In: 4th International Virtual Research Conference in Technical Disciplines, pp. 36–44 (2016)

  80. Alkhafaji, A., Fallahkhair, S., Cocea, M., Crellin, J.: A survey study to gather requirements for designing a mobile service to enhance learning from cultural heritage. In: European Conference on Technology Enhanced Learning, pp. 547–550 (2016)

  81. Asakawa, S., Guerreiro, J., Ahmetovic, D., Kitani, K.M., Asakawa, C.: The present and future of museum accessibility for people with visual impairments. In: Proceedings of the 20th International ACM SIGACCESS Conference on Computers and Accessibility, pp. 382–384 (2018)

  82. Handa, K., Dairoku, H., Toriyama, Y.: Investigation of priority needs in terms of museum service accessibility for visually impaired visitors. Br. J. Visual Impair. 28(3), 221–234 (2010)

    Google Scholar 

  83. the I. C. O. ISO Central Secretariat and IEEE, “ISO/IEC/IEEE 29148: International Standard: Systems and Software Engineering—Life Cycle Processes—Requirements Engineering (2011)

  84. Olphert, C., Damodaran, L.: Getting what you want, or wanting what you get?—Beyond user centred design. Des. Emot, pp. 138 (2004)

  85. Klein, G., Kaempf, G.L., Wolf, S., Thorsden, M., Miller, T.: Applying decision requirements to user-centered design. Int. J. Hum. Comput. Stud. 46(1), 1–15 (1997)

    Google Scholar 

  86. Maguire, M., Bevan, N.: User requirements analysis. In: IFIP World Computer Congress, TC 13, pp. 133–148 (2002)

  87. Lee, C.-H., Su, Y.-C., Chen, L.-G.: An intelligent depth-based obstacle detection system for visually-impaired aid applications. In: 2012 13th International Workshop on Image Analysis for Multimedia Interactive Services, pp. 1–4 (2012)

  88. Wei, Z., Tang, K.L., Ngan, K.N.: Implementation of H. 264 on mobile device. IEEE Trans. Consum. Electron. 53(3), 1109–1116 (2007)

    Google Scholar 

  89. Dimas, G., Ntakolia, C., Iakovidis, D.K.: Obstacle detection based on generative adversarial networks and fuzzy sets for computer-assisted navigation. In: International Conference on Engineering Applications of Neural Networks, pp. 533–544 (2019)

  90. Diamantis, D.E., Koutsiou, D.-C.C., Iakovidis, D.K.: Staircase detection using a lightweight look-behind fully convolutional neural network. In: International Conference on Engineering Applications of Neural Networks, pp. 522–532 (2019)

Download references

Acknowledgements

This research has been co-financed by the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH—CREATE—INNOVATE (Project Code: T1EDK-02070). We would like to sincerely thank the NFB for the interview and the advice, as well as all the volunteers who participated in the study. We would also like to thank all the collaborators of the ENORASI project for their contributions.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Charis Ntakolia.

Ethics declarations

Conflict of interest

On behalf of all authors, the corresponding author states that there is no conflict of interest.

Ethical approval and consent for Publication

All procedures performed in this study involving human participants were in accordance with the ethical standards of the ethics committee of the Department of Computer Science and Biomedical Informatics of University of Thessaly. Informed consent was obtained from all individual participants included in this study.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

In Table 11, we present a summary of the requirements that have been identified in the literature review regarding smart assistive systems for visual impaired individuals.

Table 11 Summary of user requirements derived from the literature

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ntakolia, C., Dimas, G. & Iakovidis, D.K. User-centered system design for assisted navigation of visually impaired individuals in outdoor cultural environments. Univ Access Inf Soc 21, 249–274 (2022). https://doi.org/10.1007/s10209-020-00764-1

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10209-020-00764-1

Keywords

Navigation