Abstract
Affect-adaptive systems mutate their behavior according to the user’s affective state. In many cases, such affective state is to be detected in a non-obtrusive way, i.e. through sensing that does not require the user to provide the system explicit input, e.g., video sensors. However, user affect recognition from video is frequently tuned to detect instantaneous emotional states, rather than longer term and more constant affective states such as mood. In this paper, we propose a non-linear computational model for bridging the gap between the recognized emotions of a person captured by a video and the overall mood of the person. For the experimental validation, emotions and mood are human annotations on an affective visual database that we created on purpose. Based on features describing peculiarities and changes in the user’s emotional state, our system is able to predict the corresponding mood well above chance and more accurately than existing models.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Bradley, M.M., Peter, J.L.: Measuring emotion: the self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 25(1), 49–59 (1994)
Broekens, J., Brinkman, W.P.: AffectButton: towards a standard for dynamic affective user feedback. In: International Conference on Affective Computing & Intelligent Interaction (2009)
Cohen, I., et al.: Facial expression recognition from video sequences: temporal and static modeling. Comput. Vis. Image Underst. 91(1), 160–187 (2003)
Cowie, R., McKeown, G.: Statistical analysis of data from initial labelled database and recommendations for an economical coding scheme (2006)
Cowie, R., Sawey, M.: GTrace-General trace program from Queen’s Belfast (2011)
Douglas-Cowie, E., et al.: The HUMAINE database: addressing the collection and annotation of naturalistic and induced emotional data. In: Paiva, A., Prada, R., Picard, R.W. (eds.) ACII 2007. LNCS, vol. 4738, pp. 488–500. Springer, Heidelberg (2007). doi:10.1007/978-3-540-74889-2_43
Ekman, P.: An argument for basic emotions. Cogn. Emot. 6(3-4), 169–200 (1992)
Ekman, P.: Basic emotions. Handb. Cogn. Emot. 98, 45–60 (1999)
Gebhard, P.: ALMA – a layered model of affect. In: International Conference on Artificial Intelligent and Multi-Agent Systems (AAMAS) (2005)
Gunes, H., Piccardi, M.: Bi-modal emotion recognition from expressive face and body gestures. J. Netw. Comput. Appl. 30, 1334–1345 (2007)
Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70(1), 489–501 (2006)
Huang, G.-B., Zhu, Q.-Y., Siew, C.-K.: Extreme learning machine: a new learning scheme of feedforward neural networks. In: IEEE International Joint Conference on Neural Networks (2004)
Jenkins, J.M., Oatley, K., Stein, N.L.: Human Emotions: A Reader. Blackwell, Oxford (1998)
Katsimerou, C., Redi, J.A., Heynderickx, I.: A computational model for mood recognition. In: Dimitrova, V., et al. (eds.) UMAP 2014. LNCS, vol. 8538, pp. 122–133. Springer, Heidelberg (2014). doi:10.1007/978-3-319-08786-3_11
Kleinsmith, A., Bianchi-Berthouze, N.: Affective body expression perception and recognition: a survey. Trans. Affect. Comput. 4, 15–33 (2013)
Lane, A.M., Terry, P.C.: The nature of mood: development of a conceptual model with a focus on depression. J. Appl. Sport Psychol. 12(1), 16–33 (2000)
McKeown, G., Valstar, M.F., Cowie, R., Pantic, M.: The SEMAINE corpus of emotionally coloured character interactions. In: IEEE International Conference on Multimedia and Expo (ICME) (2010)
Mehrabian, A.: Pleasure-arousal-dominance: a general framework for describing and measuring individual differences in temperament. Curr. Psychol. 14, 261–292 (1996)
Metallinou, A., Katsamanis, A., Narayanan, S.: Tracking continuous emotional trends of participants during affective dyadic interactions using body language and speech information. Image Vis. Comput. 31(2), 137–152 (2013)
Metallinou, A., Narayanan, S.: Annotation and processing of continuous emotional attributes: challenges and opportunities. In: IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG) (2013)
Metallinou, A., et al.: Context-sensitive learning for enhanced audiovisual emotion classification. IEEE Trans. Affect. Comput. 3(2), 184–198 (2012)
Morris, W.N.: Some thoughts about mood and its regulation. Psychol. Inq. 11, 200–202 (2000)
Nicolaou, M., Gunes, H., Pantic, M.: Continuous prediction of spontaneous affect from multiple cues and modalities in valence-arousal space. IEEE Trans. Affect. Comput. 2, 92–105 (2011)
Nicolaou, M.A., Pavlovic, V., Pantic, M.: Dynamic probabilistic CCA for analysis of affective behaviour. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7578, pp. 98–111. Springer, Heidelberg (2012). doi:10.1007/978-3-642-33786-4_8
Oatley, K., Johnson-Laird, P.N.: Towards a cognitive theory of emotions. Cogn. Emot. 1(1), 29–50 (1987)
Rusell, J.: A circumplex model of affect. Pers. Soc. Psychol. 39, 1161–1178 (1980)
Russell, J.A.: Core affect and the psychological construction of emotion. Psychol. Rev. 110(1), 145 (2003)
Sigal, L., Fleet, D.J., Troje, N.F., Livne, M.: Human attributes from 3D pose tracking. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6313, pp. 243–257. Springer, Heidelberg (2010). doi:10.1007/978-3-642-15558-1_18
Thayer, R.E.: The Origin of Everyday Moods: Managing Energy, Tension, and Stress. Oxford University Press, Oxford (1996)
Thrasher, M., Zwaag, M.D., Bianchi-Berthouze, N., Westerink, J.H.D.M.: Mood recognition based on upper body posture and movement features. In: D’Mello, S., Graesser, A., Schuller, B., Martin, J.-C. (eds.) ACII 2011. LNCS, vol. 6974, pp. 377–386. Springer, Heidelberg (2011). doi:10.1007/978-3-642-24600-5_41
Valstar, M., et al.: AVEC 2013: the continuous audio/visual emotion and depression recognition challenge. In: 3rd ACM International Workshop on Audio/Visual Emotion Challenge (2013)
Västfjäll, D.: Emotion induction through music: a review of the musical mood induction procedure. Music Sci. 5(1), 173–211 (2002)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Katsimerou, C., Redi, J.A. (2016). Neural Prediction of the User’s Mood from Visual Input. In: Baldoni, M., et al. Principles and Practice of Multi-Agent Systems. CMNA IWEC IWEC 2015 2015 2014. Lecture Notes in Computer Science(), vol 9935. Springer, Cham. https://doi.org/10.1007/978-3-319-46218-9_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-46218-9_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46217-2
Online ISBN: 978-3-319-46218-9
eBook Packages: Computer ScienceComputer Science (R0)