skip to main content
10.1145/2666242.2666245acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article

Context in Affective Multiparty and Multimodal Interaction: Why, Which, How and Where?

Authors Info & Claims
Published:16 November 2014Publication History

ABSTRACT

Recent advances in Affective Computing (AC) include research towards automatic analysis of human emotionally enhanced behavior during multiparty interactions within different contextual settings. Current paper delves on how is context incorporated into multiparty and multimodal interaction within the AC framework. Aspects of context incorporation such as importance and motivation for context incorporation, appropriate emotional models, resources of multiparty interactions useful for context analysis, context as another modality in multimodal AC and context-aware AC systems are addressed as research questions reviewing the current state-of-the-art in the research field. Challenges that arise from the incorporation of context are identified and discussed in order to foresee future research directions in the domain. Finally, we propose a context incorporation architecture into affect-aware systems with multiparty interaction including detection and extraction of semantic context concepts, enhancing emotional models with context information and context concept representation in appraisal estimation.

References

  1. G. D. Abowd, A. K. Dey, P. J. Brown, N. Davies, M. Smith, and P. Steggles. Towards a better understanding of context and context-awareness. In Handheld and Ubiquitous Computing, pages 304--307. Springer, 1999. Google ScholarGoogle Scholar
  2. S. Al Moubayed, J. Edlund, and J. Gustafson. Analysis of gaze and speech patterns in three-party quiz game interaction. In INTERSPEECH, pages 1126--1130, 2013.Google ScholarGoogle Scholar
  3. S. O. Ba and J.-M. Odobez. Multi-party focus of attention recognition in meetings from head pose and multimodal contextual cues. In International Conference on Acoustics, Speech and Signal Processing,ICASSP, pages 2221--2224. IEEE, 2008.Google ScholarGoogle ScholarCross RefCross Ref
  4. S. O. Ba and J.-M. Odobez. Recognizing visual focus of attention from head pose in natural meetings. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 39(1):16--33, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. R. Bock, A. Wendemuth, S. Gluge, and I. Siegert. Annotation and classification of changes of involvement in group conversation. In Proceedings of the Humaine Association Conference on Affective Computing and Intelligent Interaction, pages 803--808. IEEE, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. F. Bonin, R. Bock, and N. Campbell. How do we react to context? annotation of individual and group engagement in a video corpus. In Privacy, Security, Risk and Trust (PASSAT), International Conference on Social Computing (SocialCom), pages 899--903. IEEE, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. D. Borth, R. Ji, T. Chen, T. Breuel, and S.-F. Chang. Large-scale visual sentiment ontology and detectors using adjective noun pairs. In Proceedings of the 21st ACM International Conference on Multimedia, pages 223--232. ACM, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. P. J. Brown, J. D. Bovey, and X. Chen. Context-aware applications: from the laboratory to the marketplace. Personal Communications, 4(5):58--64, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  9. R. A. Calvo and S. D'Mello. Affect detection: An interdisciplinary review of models, methods, and their applications. IEEE Transactions on Affective Computing, 1(1):18--37, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. J. M. Carroll and J. A. Russell. Do facial expressions signal specific emotions? judging emotion from the face in context. Journal of Personality and Social Psychology, 70(2):205, 1996.Google ScholarGoogle ScholarCross RefCross Ref
  11. T. Choudhury and A. Pentland. Modeling face-to-face communication using the sociometer. 5:3--8, 2003.Google ScholarGoogle Scholar
  12. J. R. Curhan and A. Pentland. Thin slices of negotiation: predicting outcomes from conversational dynamics within the first 5 minutes. Journal of Applied Psychology, 92(3):802, 2007.Google ScholarGoogle ScholarCross RefCross Ref
  13. A. Dhall, R. Goecke, S. Lucey, and T. Gedeon. Collecting large, richly annotated facial-expression databases from movies. IEEE MultiMedia, 19(3):0034, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Z. Duric, W. D. Gray, R. Heishman, F. Li,A. Rosenfeld, M. J. Schoelles, C. Schunn, and H. Wechsler. Integrating perceptual and cognitive modeling for adaptive and intelligent human-computer interaction. Proceedings of the IEEE, 90(7):1272--1289, 2002.Google ScholarGoogle ScholarCross RefCross Ref
  15. {15} S. Favre, H. Salamin, J. Dines, and A. Vinciarelli. Role recognition in multiparty recordings using social affiliation networks and discrete distributions. In Proceedings of the 10th international conference on Multimodal interfaces, pages 29--36. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. D. Gatica-Perez. Analyzing group interactions in conversations: a review. In International Conference on Multisensor Fusion and Integration for Intelligent Systems, pages 41--46. IEEE, 2006.Google ScholarGoogle ScholarCross RefCross Ref
  17. D. Gatica-Perez. Automatic nonverbal analysis of social interaction in small groups: A review. Image and Vision Computing, 27(12):1775--1787, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. H. Gunes and B. Schuller. Categorical and dimensional affect analysis in continuous input: Current trends and future directions. Image and Vision Computing, 31(2):120--136, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. N. Jovanovic and R. op den Akker. Towards automatic addressee identification in multi-party dialogues. Association for Computational Linguistics, 2004.Google ScholarGoogle Scholar
  20. M. Koutsombogera, S. A. Moubayed, B. Bollepalli, A. H. Abdelaziz, M. Johansson, J. D. Á. Lopes, J. Novikova, C. Oertel, K. Stefanov, and G. Varol. The tutorbot corpus; a corpus for studying tutoring behaviour in multiparty face-to-face spoken dialogue. In LREC, pages 4196--4201, 2014.Google ScholarGoogle Scholar
  21. D. Lenat. The dimensions of context-space. Cycorp Technical Report, 1998.Google ScholarGoogle Scholar
  22. H. Liu and P. Singh. Conceptnet-a practical commonsense reasoning toolkit. BT technology journal, 22(4):211--226, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. T. Masuda, P. C. Ellsworth, B. Mesquita, J. Leu, S. Tanida, and E. Van de Veerdonk. Placing the face in context: cultural differences in the perception of facial emotion. Journal of personality and social psychology, 94(3):365, 2008.Google ScholarGoogle Scholar
  24. G. A. Miller. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39--41, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. M. Mortillaro, B. Meuleman, and K. R. Scherer. Advocating a componential appraisal model to guide emotion recognition. International Journal of Synthetic Emotions (IJSE), 3(1):18--32, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. M. Pantic, A. Nijholt, A. Pentland, and T. S. Huanag. Human-centred intelligent human-computer interaction (hci2): how far are we from attaining it? International Journal of Autonomous and Adaptive Communications Systems, 1(2):168--187, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. F. Pianesi, N. Mana, A. Cappelletti, B. Lepri, and M. Zancanaro. Multimodal recognition of personality traits in social interactions. In Proceedings of the 10th international conference on Multimodal interfaces, pages 53--60. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. N. S. Ryan, J. Pascoe, and D. R. Morse. Enhanced reality fieldwork: the context-aware archaeological assistant. In V. Gaffney, M. van Leusen, and S. Exxon, editors, Computer applications in archaeology, British Archaeological Reports, pages 182--196. Tempus Reparatum, 1998.Google ScholarGoogle Scholar
  29. H. Salamin and A. Vinciarelli. Automatic role recognition in multiparty conversations: An approach based on turn organization, prosody, and conditional random fields. IEEE Transactions on Multimedia, 14(2):338--345, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. B. Schilit, N. Adams, and R. Want. Context-aware computing applications. In Proceedings of the First Workshop on Mobile Computing Systems and Applications, pages 85--90. IEEE, 1994. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. M. Soleymani, M. Larson, T. Pun, and A. Hanjalic. Corpus development for affective video indexing. IEEE Transactions on Multimedia, 16(4), 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. M. Soleymani and M. Pantic. Human-centered implicit tagging: Overview and perspectives. In International Conference on Systems, Man, and Cybernetics (SMC), pages 3304--3309. IEEE, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  33. P. J. Stone, D. C. Dunphy, M. S. Smith, and D. M. Ogilvie. The General Inquirer: A Computer Approach to Content Analysis. MIT Press, Cambridge, MA, 1966.Google ScholarGoogle Scholar
  34. F. Talantzis, A. Pnevmatikakis, and A. G. Constantinides. Audio-visual active speaker tracking in cluttered indoors environments. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 38(3):799--807, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. A. Vinciarelli. Speakers role recognition in multiparty audio recordings using social network analysis and duration distribution modeling. IEEE Transactions on Multimedia, 9(6):1215--1226, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. A. Vinciarelli, A. Dielmann, S. Favre, and H. Salamin. Canal9: A database of political debates for analysis of social interactions. In Proceedings of the 3rd International Conference on Affective Computing and Intelligent Interaction,(ACII), pages 1--4. IEEE, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  37. A. Vinciarelli, H. Salamin, and M. Pantic. Social signal processing: Understanding social interactions through nonverbal behavior analysis. In Proceedings of International Workshop on Computer Vision and Pattern Recognition for Human Behavior Workshops, pages 42--49. IEEE, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  38. M. Wöllmer, F. Eyben, B. W. Schuller, and G. Rigoll. Temporal and situational context modeling for improved dominance recognition in meetings. In Proc. of Interspeech 2012, Portland, Oregon, USA, pages 350--353. ISCA, 2012.Google ScholarGoogle Scholar
  39. B. Wrede and E. Shriberg. The relationship between dialogue acts and hot spots in meetings. In Proc. IEEE Automatic Speech Recognition and Understanding Workshop, (ASRU'03), pages 180--185, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  40. Z. Zeng, M. Pantic, G. I. Roisman, and T. S. Huang. A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(1):39--58, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. A. Zimmermann, A. Lorenz, and R. Oppermann. An operational definition of context. In Modeling an Using Context, pages 558--571. Springer Berlin Heidelberg, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Context in Affective Multiparty and Multimodal Interaction: Why, Which, How and Where?

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            UM3I '14: Proceedings of the 2014 workshop on Understanding and Modeling Multiparty, Multimodal Interactions
            November 2014
            58 pages
            ISBN:9781450306522
            DOI:10.1145/2666242

            Copyright © 2014 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 16 November 2014

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            UM3I '14 Paper Acceptance Rate8of8submissions,100%Overall Acceptance Rate8of8submissions,100%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader