Abstract
We describe a method for dynamic emotion recognition from facial expression sequences. Our model is based on learning a latent space using the Gaussian Process Latent Variable Model (GP-LVM), encapsulating facial landmarks shapes which describe a given facial expression. We incorporate the dynamic model by learning the latent representation, with the aim to respect the data’s dynamics (facial shapes should maintain their correspondence along time). Then, a Gaussian process classifier is implemented to evaluate the relevance of the latent space features in the emotion recognition task. The results show that the proposed method can efficiently model a dynamic facial emotion and recognize with high accuracy a facial emotion sequence.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Ekman, P.: Emotions Revealed: Recognizing Faces and Feelings to Improve Communication and Emotional Life. 2nd edn. Owl Books, 175 Fifth Avenue, New York (2007)
Pantic, M., Rothkrantz, L.J.M.: Toward an affect-sensitive multimodal human-computer interaction. Proceedings of the IEEE, 1370–1390 (2003)
Zeng, Z., Pantic, M., Roisman, G., Huang, T.: A survey of affect recognition methods: Audio, visual and spontaneous expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence 31, 39–58 (2009)
Valstar, M.F., Pantic, M.: Fully automatic recognition of the temporal phases of facial actions. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 42(1), 28–43 (2012)
Chakraborty, A., Konar, A., Chakraborty, U.K., Chatterjee, A.: Emotion recognition from facial expressions and its control using fuzzy logic. Trans. Sys. Man Cyber. Part A 39, 726–743 (2009)
Cheon, Y., Kim, D.: Natural facial expression recognition using differential-aam and manifold learning. Pattern Recogn. 42, 1340–1350 (2009)
Gunes, H., Pantic, M.: Dimensional emotion prediction from spontaneous head gestures for interaction with sensitive artificial listeners. In: Allbeck, J., Badler, N., Bickmore, T., Pelachaud, C., Safonova, A. (eds.) IVA 2010. LNCS, vol. 6356, pp. 371–377. Springer, Heidelberg (2010)
Pantic, M., Patras, I.: Detecting facial actions and their temporal segments in nearly frontal-view face image sequences. In: Proc. IEEE Int’l Conf. on Systems, Man and Cybernetics, pp. 3358–3363 (2005)
Sminchisescu, C., Jepson, A.D.: Generative modeling for continuous non-linearly embedded visual inference. In: Brodley, C.E. (ed.) ICML. ACM International Conference Proceeding Series, vol. 69. ACM (2004)
Hou, S., Galata, A., Caillette, F., Thacker, N.A., Bromiley, P.A.: Real-time body tracking using a gaussian process latent variable model. In: ICCV, pp. 1–8. IEEE (2007)
Markov, K., Matsui, T.: Music genre classification using gaussian process models. In: 2013 IEEE International Workshop on Machine Learning for Signal Processing (MLSP), pp. 1–6 (2013)
Rudovic, O., Pantic, M., Patras, I.: Coupled gaussian processes for pose-invariant facial expression recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 35, 1357–1369 (2013)
Lawrence, N.: Probabilistic non-linear principal component analysis with gaussian process latent variable models. J. Mach. Learn. Res. 6, 1783–1816 (2005)
Ek, C.H., Torr, P., Lawrence, N.D.: Gaussian process latent variable models for human pose estimation. In: Popescu-Belis, A., Renals, S., Bourlard, H. (eds.) MLMI 2007. LNCS, vol. 4892, pp. 132–143. Springer, Heidelberg (2008)
Eleftheriadis, S., Rudovic, O., Pantic, M.: Shared gaussian process latent variable model for multi-view facial expression recognition. In: Bebis, G., et al. (eds.) ISVC 2013, Part I. LNCS, vol. 8033, pp. 527–538. Springer, Heidelberg (2013)
Lawrence, N.D., Quiñonero Candela, J.: Local distance preservation in the gp-lvm through back constraints. In: Proceedings of the 23rd International Conference on Machine Learning, ICML 2006, pp. 513–520. ACM, New York (2006)
Wang, J.M., Fleet, D.J., Hertzmann, A.: Gaussian process dynamical models. In: NIPS (2005)
Lucey, P., Cohn, J., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 94–101 (2010)
Ekman, P., Rosenberg, E.: What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS). Oxford Univ. Press (2005)
Zhou, F., De la Torre Frade, F.: Generalized time warping for multi-modal alignment of human motion. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2012)
Wang, J.M., Fleet, D.J., Member, S., Hertzmann, A.: Gaussian process dynamical models for human motion. IEEE Trans. Pattern Anal. Machine Intell. (2007)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
García, H.F., Álvarez, M.A., Orozco, Á. (2014). Gaussian Process Dynamical Models for Emotion Recognition. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2014. Lecture Notes in Computer Science, vol 8888. Springer, Cham. https://doi.org/10.1007/978-3-319-14364-4_77
Download citation
DOI: https://doi.org/10.1007/978-3-319-14364-4_77
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-14363-7
Online ISBN: 978-3-319-14364-4
eBook Packages: Computer ScienceComputer Science (R0)