Summary
Do users show emotions and gestures if they interact with a rather intelligent multimodal dialogue system? And if they do, what do the “emotions” and the gestures look like? Are there any features that can be exploited for their automatic detection? And finally, which language do they use when interacting with a multimodal system — does it differ from the usage of language with a monomodal dialogue system that can only understand speech?
To answer these questions, data had to be collected, labeled and analyzed. This chapter deals with the second step, the transliteration and the labeling.
The three main labeling steps are covered: orthographic transliteration, labeling of user states, labeling of gestures. Each step will be described with theoretical and developmental background, an overview of the label categories, and some practical advice for readers who are themselves in the process of looking for or assembling a coding system. Readers who are interested in using the presented labeling schemes should refer to the cited literature — not all details necessary for actually using the different systems are presented here for reasons of space. For information on the corpus itself, please refer to Schiel and Türk (2006).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
A. Batliner, K. Fischer, R. Huber, J. Spilker, and E. Nöth. Desperately Seeking Emotions Or: Actors, Wizards, and Human Beings. In: Proc. ISCA Workshop on Speech and Emotion, Belfast, Irland, 2000.
N. Beringer, S. Burger, and D. Oppermann. Lexikon der Transliterationen. Technical Document SmartKom 2, Bavarian Archive for Speech Signals (BAS), 2000.
R.R. Cornelius. Theoretical Approaches to Emotion. In: Proc. ISCA Workshop on Speech And Emotion, Belfast, Ireland, 2000.
J. Egan. Signal Detection Theory and ROC Analysis. Academic, New York, 1975.
P. Ekman. Emotional and Conversational Nonverbal Signals. In: L.S. Messing and R. Campbell (eds.), Gesture, Speech and Sign, pp. 45–55. Oxford University Press, 1999.
P. Ekman and W.V. Friesen. Facial Action Coding System (FACS). A Technique for the Measurement of Facial Action. Consulting Psychologists Press, 1978.
G. Faßnacht. Systematische Verhaltensbeobachtung. Reinhardt, Munich, Germany, 1979.
K. Fischer. Annotating Emotional Language Data. Verbmobil Report 236, DFKI, 1999.
W.V. Friesen, P. Ekman, and H.G. Wallbott. Measuring Hand Movements. Journal of Nonverbal Behavior, 1:97–112, 1979.
I. Jacobi and F. Schiel. Interlabeller Agreement in SmartKom Multi-Modal Annotations. Technical Document SmartKom 26, Bavarian Archive for Speech Signals (BAS), 2003.
A. Kendon. Gesticulation and Speech: Two Aspects of the Process. In: M.R. Key (ed.), The Relation Between Verbal and Nonverbal Communication, The Hague, The Netherlands, 1980. Mouton.
R. Laban. Principles of Dance and Movement Notation. Macdonald & Evans Ltd., London, UK, 1956.
D. McNeill. Hand and Mind: What Gestures Reveal About Thought. University of Chicago Press, Chicago, IL, 1992.
A. Mulder. Human Movement Tracking Technology. Hand Centered Studies of Human Movement Project 94-1, School of Kinesiology, Simon Fraser University, Burnaby, Canada, 1994.
N. Munn. The Effect of Knowledge of the Situation Upon Judgment of Emotion From Facial Expressions. Journal of Abnormal and Social Psychology, 35:324–338, 1940.
D. Oppermann, F. Schiel, S. Steininger, and N. Beringer. Off-Talk-A Problem for Human-Machine-Interaction. In: Proc. EUROSPEECH-01, Aalborg, Denmark, 2001.
R. Plutchik. Emotion: A Psycho-Evolutionary Synthesis. Harper and Row, New York, 1980.
C. Ratner. A Social Constructionist Critique of Naturalistic Theories of Emotion. Journal of Mind and Behavior, 10:211–230, 1989.
F. Schiel and U. Türk. Wizard-of-Oz Recordings, 2006. In this volume.
M. Sherman. The Differentiation of Emotional Responses in Infants. Journal of Comparative Physiology, pp. 265–284, 1927.
S. Steininger, O. Dioubina, R. Siepmann, C. Beiras-Cunqueiro, and A. Glesner. Labeling von User-State im Mensch-Maschine Dialog — User-State-Kodierkonventionen SmartKom. Technical Document SmartKom 17, Bavarian Archive for Speech Signals (BAS), 2000.
S. Steininger, B. Lindemann, and T. Paetzold. Labeling of Gestures in SmartKom-The Coding System. In: I. Wachsmuth and T. Sowa (eds.), Gesture and Sign Languages in Human-Computer Interaction, Int. Gesture Workshop 2001 London, pp. 215–227, Berlin Heidelberg New York, 2001a. Springer.
S. Steininger, B. Lindemann, and T. Paetzold. Labeling von Gesten im Mensch-Maschine Dialog — Gesten-Kodierkonventionen SmartKom. Technical Document SmartKom 14, Bavarian Archive for Speech Signals (BAS), 2001b.
S. Steininger, F. Schiel, O. Dioubina, and S. Rabold. Development of User-State Conventions for the Multimodal Corpus in SmartKom. In: Proc. Workshop “Multimodal Resources and Multimodal Systems Evaluation”, pp. 33–37, Las Palmas, Spain, 2002a.
S. Steininger, F. Schiel, and A. Glesner. Labeling Procedures for the Multi-Modal Data Collection of SmartKom. In: Proc. 3rd Int. Conf. on Language Resources and Evaluation (LREC 2002), pp. 371–377, Las Palmas, Spain, 2002b.
H. Wallbott. Faces in Context: The Relative Importance of Facial Expression and Context Information in Determining Emotion Attributions. In: K. Scherer (ed.), Facets of Emotion, Mahwah, NJ, 1988. Lawrence Erlbaum.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Steininger, S., Schiel, F., Rabold, S. (2006). Annotation of Multimodal Data. In: Wahlster, W. (eds) SmartKom: Foundations of Multimodal Dialogue Systems. Cognitive Technologies. Springer, Berlin, Heidelberg . https://doi.org/10.1007/3-540-36678-4_35
Download citation
DOI: https://doi.org/10.1007/3-540-36678-4_35
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-23732-7
Online ISBN: 978-3-540-36678-2
eBook Packages: Computer ScienceComputer Science (R0)