Abstract
This paper explores the theoretical aspect of providing context to contextually impaired individuals and offers some considerations on how a machine learning system can be adapted to learn contextual clues and then provide these to users. Context itself, is the clarifying component of a situation and helps people to understand what is happening and why. Some people struggle to understand the situations that they are in, due to several issues, particularly those with sensory or social impairments, such as autism. These impairments can be interpreted as form of contextual blindness, where a lack of awareness of certain contextual clues render communication or understanding of it difficult. These clues can be as simple as a person’s current location, to the more nuanced impact of a conversational partner’s body language. Since awareness of the current context of a situation can help to clarify its meaning, it is expected that provision of contextual information to contextually impaired individuals will be beneficial; by giving them a point of reference for the tone of the situation they are in and supporting them to react accordingly. The results presented demonstrate the impact of contextual information in a specific social situation (dating), giving an indication of the benefits of both context and contextual awareness, which provide the basis for further investigation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Moraes, A.d.: First Dates. Twenty Twenty/Channel 4, UK, First aired (2013)
Kinch, M.W., Melis, W.J., Keates, S.: The benefits of contextual information for speech recognition systems. In: 10th Computer Science and Electronic Engineering (CEEC), Colchester, UK (2018)
Smith, S.M., Vela, E.: Environmental context-dependent memory: a review and meta-analysis. Psychon. Bull. Rev. 8, 203–220 (2001)
Smith, S.M.: Background music and context-dependent memory. Am. J. Psychol. 98(4), 591–603 (1985)
Ng, S., Payne, B.R., Stine-Morrow, E.A., Federmeier, K.D.: How struggling adult readers use contextual information when comprehending speech: evidence from event-related potentials. Int. J. Psychophysiol. 125, 1–9 (2018)
Godden, D.R., Baddeley, A.D.: Context-dependent memory in two natural environments: on land and underwater. Br. J. Psychol. 66, 325–331 (1975)
Dey, K., Abowd, G.D.: Towards a better understanding of context and context-awareness. In: Proceedings of the 1st International Symposium on Handheld and Ubiquitous Computing (1999)
Ryan, N.S., Pascoe, J., Morse, D.R.: Enhanced reality fieldwork: the context-aware archaeological assistant. In: Computer Applications in Archaeology (1998)
Kinch, M.W., Melis, W.J., Keates, S.: Reviewing the current state of machine learning for artificial intelligence with regards to the use of contextual information. In: The Second Medway Engineering Conference on Systems: Efficiency, Sustainablity and Modelling, Medway, UK (2017)
Want, R., Hopper, A., Falcao, V., Gibbons, J.: The active badge location system. ACM Trans. Inform. Syst. 10, 91–102 (1992)
Williams, I., Kannan, A., Aleksic, P., Rybach, D., Sainath, T.N., Contextual speech recognition in end-to-end neural network systems using beam search. In: Interspeech 2018 (2018)
Baron-Cohen, S.: Mindblindness: An Essay on Autism and Theory of Mind. The MIT Press, Cambridge, USA (1995)
Baron-Cohen, S., Wheelwright, S., Hill, J., Raste, Y., Plumb, I.: The “Reading the Mind in the Eyes” test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism. J. Child Psychol. Psychiatry 42(2), 241–251 (2001)
Mongeau, P.A., Serewicz, M.C.M., Therrien, L.F.: Goals for cross-sex first dates: identification, measurement, and the influence of contextual factors. Commun. Monogr. 71(2), 121–147 (2004)
Argyle, M.: Bodily Communication. Routledge, New York, USA (2013)
Ekman, P., Sorenson, E.R., Friesen, W.V.: Pan-cultural elements in facial displays of emotion. Science 164, 86–88 (1969)
Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. J. Pers. Soc. Psychol. 17(2), 124–129 (1971)
Darwin, C.: The Expression of the Emotions in Man and Animals. John Murray, London (1872)
McDuff, D., Mahmoud, A., Mavadati, M., Amr, M., Turcot, J., El Kaliouby, R.: AFFDEX SDK: a cross-platform real-time multi-face expression recognition toolkit. In: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, New York, USA (2016)
Rudovic, O., Lee, J., Dai, M., Schuller, B., Picard, R.W.: Personalized machine learning for robot perception of affect and engagement in autism therapy. Sci. Robot. 3(19), 1–11 (2018)
El Kaliouby, R., Robinson P., Keates, S.: Temporal context and the recognition of emotion from facial expression. In: HCI International, Crete (2003)
El Kaliouby, R., Robinson, P.: The emotional hearing aid - an assistive tool for autism. In: HCI International, Crete (2003)
Channel 4, All 4. https://www.channel4.com/
Learning on Screen. https://learningonscreen.ac.uk/ondemand
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Kinch, M., Keates, S. (2021). Contextual Cues: The Role of Machine Learning in Supporting Contextually Impaired Users. In: Antona, M., Stephanidis, C. (eds) Universal Access in Human-Computer Interaction. Design Methods and User Experience. HCII 2021. Lecture Notes in Computer Science(), vol 12768. Springer, Cham. https://doi.org/10.1007/978-3-030-78092-0_39
Download citation
DOI: https://doi.org/10.1007/978-3-030-78092-0_39
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-78091-3
Online ISBN: 978-3-030-78092-0
eBook Packages: Computer ScienceComputer Science (R0)