Skip to main content

Towards Multimodal Affective Stimulation: Interaction Between Visual, Auditory and Haptic Modalities

  • Conference paper
  • First Online:
ICT Innovations 2016 (ICT Innovations 2016)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 665))

Included in the following conference series:

Abstract

Affective computing is concerned with designing and implementing emotionally intelligent machines. Three major subareas of research in this field are: (1) sensing the emotional state of the user, (2) expressing or displaying emotional states in a robot or an avatar, and (3) manipulating the users emotional state. For example, Picard [1] has designed a system to infer emotional state of the user based on their facial expressions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Picard, R.W.: Future affective technology for autism and emotion communication. Philos. Trans. Roy. Soc. B. Biol. Sci. 364(1535), 3575–3584 (2009)

    Article  Google Scholar 

  2. Lee, C.H.J., Chang, C., Chung, H., Dickie, C., Selker, T.: Emotionally reactive television. In: Proceedings of the 12th International Conference on Intelligent User Interfaces, pp. 329–332. ACM (2007)

    Google Scholar 

  3. Danziger, S., Levav, J., Avnaim-Pesso, L.: Extraneous factors in judicial decisions. Proc. Nat. Acad. Sci. 108(17), 6889–6892 (2011)

    Article  Google Scholar 

  4. Jayant, C., Acuario, C., Johnson, W., Hollier, J., Ladner, R.: V-braille: haptic braille perception using a touch-screen and vibration on mobile phones. In: Proceedings of the 12th international ACM SIGACCESS conference on Computers and accessibility, pp. 295–296. ACM (2010)

    Google Scholar 

  5. Amemiya, T., Sugiyama, H.: Design of a haptic direction indicator for visually impaired people in emergency situations. Comput. Helping People Spec. Needs pp. 1141–1144 (2008)

    Google Scholar 

  6. Tsetserukou, D., Neviarouskaya, A., Prendinger, H., Ishizuka, M., Tachi, S.: iFeel_IM: innovative real-time communication system with rich emotional and haptic channels. In: CHI 2010 Extended Abstracts on Human Factors in Computing Systems, pp. 3031–3036. ACM (2010)

    Google Scholar 

  7. McGurk, H., MacDonald, J.: Hearing lips and seeing voices (1976)

    Google Scholar 

  8. Alais, D., Burr, D.: The ventriloquist effect results from near-optimal bimodal integration. Curr. Biol. 14(3), 257–262 (2004)

    Article  Google Scholar 

  9. Yokosawa, K., Kanaya, S.: Ventriloquism and audio-visual integration of voice and face. Brain and Nerve= Shinkei Kenkyu no Shinpo 64(7), 771–777 (2012)

    Google Scholar 

  10. Botvinick, M., Cohen, J.: Rubber hands’ feel’touch that eyes see. Nature 391(6669), 756 (1998)

    Article  Google Scholar 

  11. Jousmäki, V., Hari, R.: Parchment-skin illusion: sound-biased touch. Curr. Biol. 8(6), R190–R191 (1998)

    Article  Google Scholar 

  12. Ernst, M.O., Banks, M.S.: Humans integrate visual and haptic information in a statistically optimal fashion. Nature 415(6870), 429–433 (2002)

    Article  Google Scholar 

  13. Tsakiris, M., Haggard, P.: The rubber hand illusion revisited: visuotactile integration and self-attribution. J. Exp. Psychol. Hum. Percept. Perform. 31(1), 80 (2005)

    Article  Google Scholar 

  14. Fagel, S.: Emotional McGurk effect. In: Proceedings of the International Conference on Speech Prosody, Dresden, vol. 1. Citeseer (2006)

    Google Scholar 

  15. Li, A., Dang, J.: A cross-cultural investigation on emotion expression under vocal and facial conflict-also an observation on emotional McGurk effect. In: International Symposium on Biomechanical and Physiological Modeling and Speech Science, Kanazawa, I, pp. 37–50 (2009)

    Google Scholar 

  16. Alagarai Sampath, H., Indurkhya, B., Lee, E., Bae, Y., et al.: Towards multimodal affective feedback: interaction between visual and haptic modalities. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pp. 2043–2052. ACM (2015)

    Google Scholar 

  17. Larsen, R.J., Diener, E.: Promises and problems with the circumplex model of emotion (1992)

    Google Scholar 

  18. Lang, P.J., Bradley, M.M., Cuthbert, B.N.: International affective picture system (IAPS): technical manual and affective ratings. NIMH Center for the Study of Emotion and Attention, pp. 39–58 (1997)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bipin Indurkhya .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this paper

Cite this paper

Indurkhya, B. (2018). Towards Multimodal Affective Stimulation: Interaction Between Visual, Auditory and Haptic Modalities. In: Stojanov, G., Kulakov, A. (eds) ICT Innovations 2016. ICT Innovations 2016. Advances in Intelligent Systems and Computing, vol 665. Springer, Cham. https://doi.org/10.1007/978-3-319-68855-8_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-68855-8_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-68854-1

  • Online ISBN: 978-3-319-68855-8

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics