A Social-Aware Assistant to support individuals with visual impairments during social interaction: A systematic requirements analysis

https://doi.org/10.1016/j.ijhcs.2018.08.007Get rights and content

Highlights

  • A socially-aware assistant, consisting of both perceptive and feedback components.

  • Our approach is based on a wearable platform.

  • Our system has been tested with visually impaired people.

Abstract

Visual impairment affects the normal course of activities in everyday life including mobility, education, employment, and social interaction. Most of the existing technical solutions devoted to empowering the visually impaired people are in the areas of navigation (obstacle avoidance), access to printed information and object recognition. Less effort has been dedicated so far in developing solutions to support social interactions. In this paper, we introduce a Social-Aware Assistant (SAA) that provides visually impaired people with cues to enhance their face-to-face conversations. The system consists of a perceptive component (represented by smartglasses with an embedded video camera) and a feedback component (represented by a haptic belt). When the vision system detects a head nodding, the belt vibrates, thus suggesting the user to replicate (mirror) the gesture. In our experiments, sighted persons interacted with blind people wearing the SAA. We instructed the former to mirror the noddings according to the vibratory signal, while the latter interacted naturally. After the face-to-face conversation, the participants had an interview to express their experience regarding the use of this new technological assistant. With the data collected during the experiment, we have assessed quantitatively and qualitatively the device usefulness and user satisfaction.

Introduction

The visual impairment affects almost every activity of daily living, including mobility, orientation, education, employment, and social interaction. Despite these limitations, blind people reinforce their other perceptual mechanisms: hearing (Kolarik et al., 2016), tact (Voss et al., 2016) and olfact (Araneda et al., 2016), as a form to compensate for their impairment. Some blind people are even capable of developing echolocation, which is the ability to detect objects in their environment by actively creating sounds, e.g., by tapping their canes, lightly stomping their foot, snapping their fingers, or making clicking noises with their mouths (Secchi et al., 2017), and sensing echoes from sound bouncing on the objects. A skilled blind pedestrian can approach an intersection, listen to the traffic, and from audible information alone, judge the spatial layout of intersecting streets, the width of the road, the traffic volume, and the presence of pedestrian islands or medians  (Liu and Sun, 2006). Likewise, blind people tend to develop a highly sensitive tactile sense that allows them to read Braille system (Russomanno et al., 2015), and to scan their surroundings (Goldreich and Kanics, 2003). However, even though blind people tend to compensate their lack of sight by augmenting their other senses, the sensory bandwidth of the visual channel is orders of magnitude greater than that of auditory and touch  (Loomis et al., 2012).

Nowadays, we witness an unprecedented effort of the scientific community to develop solutions which could restore, even partially, the sense of sight. As a result, several assistive systems using computer vision have been developed for different applications: access to printed information  (Cutter, Manduchi, 2013, Cutter, Manduchi, 2015), object recognition  (Gallo, Manduchi, 2011, Guida, Comanducci, Colombo, 2011, Winlock, Christiansen, Belongie, 2010), navigation  (Flores, Manduchi, Zenteno, 2014, Vera, Zenteno, Salas, 2013) and social interaction  (Krishna, Bala, McDaniel, McGuire, Panchanathan, 2010, Krishna, Panchanathan, 2010). A comprehensive survey of these applications could be found in Manduchi and Coughlan (2012). Of all these areas, the least exploited one is social interaction (Loomis et al., 2012). The importance of the visual channel is reflected in the type of information exchanged with our counterparts during social interactions. According to Ambady and Rosenthal (1992), nearly 65% of the generated volume is through non-verbal cues, i.e., eye gaze, facial/hand/body gestures. Thus, a visually impaired person suffers a serious limitation in accessing this rich flow of information and in consequence this deprivation could lead to social isolation, depression, loneliness, and anxiety  (Hinds et al., 2003).

In this work, we develop a wearable technological assistant to provide visually impaired people with cues that may enhance their social interaction during face-to-face conversations. The system architecture is depicted in Fig. 1(a). It consists of a perceptive component (represented by smartglasses which have embedded a video camera shown in Fig. 1(b)) and a reactive component (represented by a haptic belt shown in Fig. 1(c)). The video stream captured by the camera is sent to a computer, where a specialized software runs face detection, tracking, and head gestures algorithms. When a head nodding is detected, this information is sent to the vibratory belt informing the user, who, on his turn, could replicate the gesture. By closing this gesture loop, a mirroring event is triggered.

The main contributions of our approach are:

  • The introduction a Socially-Aware Assistant (SAA), consisting of perceptive, and feedback components.

  • The implementation of a computer-vision-based application for a social assistant on a wearable platform. The use of smartglasses confers the system a first-person perspective, which is ideal for this type of applications when compared with other assistive systems that offer a third-person perspective.

  • The development of new insights obtained by testing the SAA with visually impaired people. We validated our system in a simulated tourism office environment, where we envisioned a face-to-face conversation taking place between a blind person (acting as a tourist guide) and a sighted person (acting as tourists). As a result, a custom data corpus has been recorded, annotated, and will be made available to the research community upon request.

Our application integrates work we have previously done on head-gesture recognition (Terven, Raducanu, Meza-de Luna, Salas, 2016, Terven, Salas, Raducanu, 2014) and social interaction (Meza-de Luna et al., 2016), and new results which include the development of a real-time technology assistant for visual impaired social interaction, and the use of a vibratory belt for haptic feedback.

The rest of the paper is structured as follows: we dedicate Section 2 to review the related work in the area of assistive technology for social interaction. In Section 3, we describe the system’s architecture. In Section 4, we explain the design of the experimental study. In Section 5, we report our results and discuss the findings of our study. Finally, we draw our conclusion and provide some guidelines for future work.

Section snippets

Related literature

To a large extent, assistive technology for visually impaired people is based on ultrasonic, infrared, or laser sensors (Dakopoulos and Bourbakis, 2010). However, as sighted people acquire most of their information through visual perception, it is tempting to use computer vision to achieve the same goal (Chessa et al., 2016). Nonetheless, limitations in computing power and the lack of reliable algorithms have been a recurrent problem for computer vision in this area. Fortunately, parallel to

System architecture

Our system’s architecture consists of a perceptual component represented by a wearable camera embedded in smartglasses (to detect the head noddings) and a feedback component represented by a haptic belt (to inform the blind user when such an event has been detected) (see Fig. 2). The pipeline for our head-nodding detection algorithm consists of the following steps: face detection, face-tracking and stabilization, head pose estimation and head gesture recognition. We explain these elements in

Experimental design

To assess the usefulness of the SAA, we designed an experiment integrating mixed methods (Hernández et al., 2010). To facilitate the conversation between two unknown people, we set up a scenario of a tourist office where the blind person could explain freely what can be visited in the city. The scene offered an accessible and motivating conversation topic for people, regardless of their level of education. To achieve this, we selected participants among the ones who knew the city of Querétaro

Experimental results

In the experiment described in Section 4, visually impaired people acted as tourist guides while normally sighted people acted as tourists (See Fig. 3). In this section, we analyze the result of the conversations. As our objective is to find out the forms the SAA improves the communication with blind people, we search for the strength of the relationship between the outcome of the inferences made with the visual cues we obtain during the conversation, and the responses of the participants to

Conclusion

This paper introduces a wearable assistant for blind users to support their social interaction. The SAA developed is capable of automatically recognizing head-noddings and to incite mirroring events using visual information obtained from cameras and providing haptic feedback to the blind user. With the inference of satisfaction derived from the qualitative evaluation of the interviews as ground truth, we showed that our computer vision algorithms are capable of recognizing whether the SAA was

Acknowledgments

This work was partially supported by CONACYT under Grant INFR-2016-01: 268542, by IPN-SIP under Grant No. 20170566 and by UCMexus, for Joaquín Salas; and by MINECO Grant TIN2016-79717-R, Spain, for Bogdan Raducanu; Juan Ramón Terven was partially supported by Instituto Tecnológico Nacional de México and CONACYT.

María Elena Meza-de-Lunais a researcher at the Universidad Autónoma de Querétaro, México. She is interested in the social and cultural issues to prevent the expression of violence. Currently, she is director of !Atrévete Ya!/iHollaback!-Querétaro, an organization to prevent street harassment (www.atrevete-ya.org) and president of IIPSIS, an NGO devoted to research and intervention in psychosocial matters. Contact her at [email protected].

References (79)

  • R. Araneda et al.

    Cortical plasticity and olfactory function in early blindness

    Front. Syst. Neurosci.

    (2016)
  • M. Auvray et al.

    There is something out there: distal attribution in sensory substitution, twenty years later

    J. Integr. Neurosci.

    (2005)
  • M. Auvray et al.

    Perception with compensatory devices: from sensory substitution to sensorimotor extension

    Cogn. Sci.

    (2009)
  • P. Bach-y Rita

    Sensory plasticity

    Acta Neurol. Scand.

    (1967)
  • S. Battiato et al.

    SIFT features tracking for video stabilization

    IEEE International Conference on Image Analysis and Processing

    (2007)
  • L. Breiman

    Random forests

    Mach. Learn.

    (2001)
  • H. Buimer et al.

    Conveying facial expressions to blind and visually impaired persons through a wearable vibrotactile device

    PLoS One

    (2018)
  • C. Capelle et al.

    A real-Time experimental prototype for enhancement of vision rehabilitation using auditory substitution

    IEEE Trans. Biomed. Eng.

    (1998)
  • C. Collins

    Tactile television-Mechanical and electrical image projection

    IEEE Trans. Man-Mach. Syst.

    (1970)
  • M. Cutter et al.

    Real time camera phone guidance for compliant document image acquisition without sight

    IEEE International Conference on Document Analysis and Recognition

    (2013)
  • M. Cutter et al.

    Towards mobile OCR: how to take a good picture of a document without sight

    International Symposium on Document Engineering

    (2015)
  • D. Dakopoulos et al.

    Wearable obstacle avoidance electronic travel aids for blind: A Survey

    IEEE Trans. Syst. Man Cybernet. Part C

    (2010)
  • D. Dementhon et al.

    Model-Based object pose in 25 lines of code

    Int. J. Comput. Vis.

    (1995)
  • B. Deville et al.

    Guiding the focus of attention of blind people with visual saliency

    Workshop on Computer Vision Applications for the Visually Impaired

    (2008)
  • S. Farley

    Nonverbal reactions to an attractive stranger: the role of mimicry in communicating preferred social distance. journal of nonverbal behavior

    J. Nonverbal. Behav.

    (2014)
  • A. Fathi et al.

    Social interactions: a first-person perspective

    IEEE Conference on Computer Vision and Pattern Recognition

    (2012)
  • R. Fitzgerald et al.

    Reactions to blindness: a four-year follow-up study

    Percept Mot Skills

    (1987)
  • L. Flavio et al.

    Publication manual of the American psychological association

    Reports, Results and Recommendations from Technical Events Series

    (2010)
  • G. Flores et al.

    Vibrotactile guidance for wayfinding of blind walkers

    IEEE Trans. Haptics

    (2015)
  • G. Flores et al.

    Ariadne’s thread: robust turn detection for path back-tracing using the iPhone

    Ubiquitous Positioning Indoor Navigation and Location Based Service

    (2014)
  • M. Frame

    Blind Spots: The Communicative Performance of Visual Impairment in Relationships and Social Interaction

    (2004)
  • O. Gallo et al.

    Reading 1D barcodes with mobile phones using deformable templates

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2011)
  • R. Gifford et al.

    Nonverbal cues in the employment interview: links between applicant qualities and interviewer judgements

    Appl. Psychol.

    (1985)
  • D. Goldreich et al.

    Tactile acuity is enhanced in blindness

    J. Neurosci.

    (2003)
  • T. Grove et al.

    Conversations with ablebodied and visibly disabled strangers: an adversarial test of predicted outcome value and uncertainty reduction theories

    Hum. Commun. Res.

    (1991)
  • M. Grundmann et al.

    Auto-directed video stabilization with Robust L1 optimal camera paths

    IEEE Conference on Computer Vision and Pattern Recognition

    (2011)
  • N. Guéguen et al.

    Mimicry in social interaction: its effect on human judgement and behavior

    Eur. J. Soc. Sci.

    (2009)
  • C. Guida et al.

    Automatic Bus Line Number Localization and Recognition on Mobile Phones. a Computer Vision Aid for the Visually Impaired

    Image Analysis and Processing

    (2011)
  • D. Hayden

    Wearable-assisted social interaction as assistive technology for the blind

    Master Thesis

    (2014)
  • Cited by (22)

    • A systematic review of intelligent assistants

      2022, Future Generation Computer Systems
      Citation Excerpt :

      Other research efforts have proposed IAs considering user privacy since their design. That is the case of [42] that provides users with a bidirectional silent speech interface and [71] that gives feedback to users utilizing a haptic device. However, both studies provide a sense of user privacy with respect to how the user interacts with the IA and not regarding what data is collected and how and where is stored.

    View all citing articles on Scopus

    María Elena Meza-de-Lunais a researcher at the Universidad Autónoma de Querétaro, México. She is interested in the social and cultural issues to prevent the expression of violence. Currently, she is director of !Atrévete Ya!/iHollaback!-Querétaro, an organization to prevent street harassment (www.atrevete-ya.org) and president of IIPSIS, an NGO devoted to research and intervention in psychosocial matters. Contact her at [email protected].

    Juan R. Terven is a senior research scientist at AiFi Inc. His research interests include computer vision, machine learning, and assistive technologies design. He received a Ph.D. in computer vision from IPN, Mexico. Contact him at [email protected].

    Bogdan Raducanuis a senior researcher and project director at the Computer Vision Center in Barcelona, Spain. His research interests include computer vision, pattern recognition, and social computing. Raducanu received a Ph.D. in computer science from the University of the Basque Country, Bilbao, Spain. Contact him at [email protected].

    Joaquín Salasis a professor in the area of Computer Vision at Instituto Politécnico Nacional. His professional activity include the development of applications based on computer vision and pattern recognition. Salas received a Ph.D. in computer science from ITESM, México. Contact him at [email protected].

    View full text