Abstract
Over the last few years, many researchers have done a lot of work on emotion recognition from facial expressions using the techniques of image processing and computer vision. In this paper we explore the application of Latent Dirichlet Allocation, a technique conventionally used in Natural text processing, when used with Hidden Markov Model, for the same. The classification is done at an image sequence level. Each frame of an image sequence is represented by a feature vector, which is mapped to one of the words from the dictionary generated using K-means. Latent Dirichlet Allocation then models each image sequence as a set of topics. We further know the order of topics for image sequence from the order of words, which we use for classification in the next step. This is done by training a Hidden Markov Model for each emotion. The emotions dealt with are six basic emotions: happy, fear, sad, surprise, angry, disgust and contempt. We compare our results with another technique in which sequence information of words instead of topics is used by HMM for learning facial expression dynamics. The results have been presented on CK+ dataset [2]. The accuracy obtained on the proposed technique is 80.77% .The use of word-sequence in found to give better results in general.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Kanade, T., Cohn, J.F., Tian, Y.: Comprehensive database for facial expression analysis. In: Proceedings of the Fourth IEEE International Conference on Automatic Face and Gesture Recognition (FG 2000), Grenoble, France, pp. 46–53 (2000)
Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The Extended Cohn-Kanade Dataset (CK+): A complete expression dataset for action unit and emotion-specified expression. In: Proceedings of the Third International Workshop on CVPR for Human Communicative Behavior Analysis (CVPR4HB 2010), San Francisco, USA, pp. 94–101 (2010)
Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion, J. Personality Social Psychol. 17(2), 124–129 (1971)
Bartlett, G., Littlewort, M., Frank, C., Lainscsek, I., Fasel, F., Movellan, J.: Automatic recognition of facial actions in spontaneous expressions. Journal of Multimedia 1(6), 22–35 (2006)
Valstar, M.F., Pantic, M.: Fully Automatic Recognition of the Temporal Phases of Facial Actions. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 42(1), 28–43 (2012)
Hong, H., Neven, H., von der Malsburg, C.: Online Facial Expression Recognition based on Personalized Galleries. In: Proc. Int’l Conf. Automatic Face and Gesture Recognition, pp. 354–359 (1998)
Huang, C.L., Huang, Y.M.: Facial Expression Recognition Using Model-Based Feature Extraction and Action Parameters Classification. J. Visual Comm. and Image Representation 8(3), 278–290 (1997)
Kimura, S., Yachida, M.: Facial Expression Recognition and Its Degree Estimation. In: Proc. Computer Vision and Pattern Recognition, pp. 295–300 (1997)
Pantic, M., Rothkrantz, L.J.: Expert System for Automatic Analysis of Facial Expression. Image and Vision Computing J. 18(11), 881–905 (2000)
Yang, P., Liu, Q., Metaxas, D.N.: Boosting coded dynamic features for facial action units and facial expression recognition. In: CVPR, pp. 1–6 (2007)
Zhao, G., Pietikainen, M.: Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. on PAMI 29(6), 915–928
Padgett, C., Cottrell, G.: Representing face image for emotion classification. In: Mozer, M., Jordan, M., Petsche, T. (eds.) Advances in Neural Information Processing Systems, vol. 9, pp. 894–900. MIT Press, Cambridge
Lanitis, A., Taylor, C., Cootes, T.: Automatic interpretation and coding of face images using flexible models. IEEE Trans. Pattern Anal. Mach. Intell. 19(7), 743–756 (1997)
Lien. Automatic recognition of facial expression using hidden Markov models and estimation of expression intensity. Ph.D. Thesis, The Robotics Institute, CMU (April 1998)
Otsuka, T., Ohya, J.: Spotting segments displaying facial expression from image sequences using HMM. In: IEEE Proceedings of the Second International Conference on Automatic Face and Gesture Recognition (FG 1998), Nara, Japan, pp. 1998–442 (1998)
Essa, I., Pentland, A.: Coding, analysis, interpretation and recognition of facial expressions. IEEE
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Bansal, A., Chaudhary, S., Roy, S.D. (2013). A Novel LDA and HMM-Based Technique for Emotion Recognition from Facial Expressions. In: Schwenker, F., Scherer, S., Morency, LP. (eds) Multimodal Pattern Recognition of Social Signals in Human-Computer-Interaction. MPRSS 2012. Lecture Notes in Computer Science(), vol 7742. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-37081-6_3
Download citation
DOI: https://doi.org/10.1007/978-3-642-37081-6_3
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-37080-9
Online ISBN: 978-3-642-37081-6
eBook Packages: Computer ScienceComputer Science (R0)