Abstract
Human emotion interpretation contributes greatly in Human-Machine Interface (HMI) spanning applications in health care, education, and entertainment. Affective interactions can have the most influence when emotional recognition is available to both human and computers. However, developing robust emotion recognizers is a challenging task in terms of modality, feature selection, and classifier and database design. Most leading research uses facial features, yet verbal communication is also fundamental for sensing affective state especially when visual information is occluded or unavailable. Recent work deploys audiovisual data in bi-modal emotion recognizers. Adding more information e.g. gesture analysis, event/scene understanding, and speaker identification, helps increase recognition accuracy. As classification of human emotions can be considered a multi-modal pattern recognition problem, in this paper, we propose the schematics of a multi-dimension system for automatic human emotion recognition.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Wu, C., Lin, J., Wei, W.: Survey on audiovisual emotion recognition: databases, features, and data fusion strategies. APSIPA Trans. Signal Inf. Process. 3, e12 (2014)
Daniel Căleanu, C.: Face expression recognition: a brief overview of the last decade. In: 2013 IEEE 8th International Symposium on Applied Computational Intelligence and Informatics (SACI), pp. 157–161, May 2013
Khatri, N.N., Shah, Z.H., Patel, S.A.: Facial expression recognition: a survey. Int. J. Comput. Sci. Inf. Technol. (IJCSIT) 5(1), 149–152 (2014)
El Ayadi, M., Kamel, M.S., Karray, F.: Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recogn. 44, 572–587 (2011)
Ragavan, S., Kittusamy, V., Chakrapani, V.: Facial expressions recognition using eigenspaces. J. Comput. Sci. 8, 1674 (2012)
Zhang, S., Zhao, X., Lei, B.: Facial expression recognition based on local binary patterns and local fisher discriminant analysis. WSEAS Trans. Sig. Process. 8, 21–31 (2012)
Elaiwat, S., Bennamoun, M., Boussaid, F., El-Sallam, A.: 3-D Face recognition using curvelet local features. Biom. Compend. 21(2), 172–175 (2014)
Drira, H., Ben Amor, B., Srivastava, A., Daoudi, M., Slama, R.: 3D face recognition under expressions, occlusions, and pose variations. Pattern Anal. Mach. Intell. 35, 2270–2283 (2013)
Mishra, B. et al.: Facial expression recognition using feature based techniques and model based techniques: a survey. In: IEEE Sponsored 2nd International Conference on Electronics and Communication System, ICECS 2015 (2015)
Hayat, M., Bennamoun, M.: An automated framework for textured 3D video based facial expression recognition. Affect. Comput. 5, 301–313 (2014)
Ekman, P., Friesen, W.: Facial Action Coding System. Consulting Psychologists Press, Washington, D.C. (1977)
Morrison, D., Wang, R., De Silva, L.C.: Ensemble methods for spoken emotion recognition in call-centers. Speech Commun. 49, 98–112 (2007)
Le Bouquin, R.: Enhancement of noisy speech signals: application to mobile radio communications. Speech Commun. 18(1), 3–19 (1996)
Bou-Ghazale, S., Hansen, J.: A comparative study of traditional and newly proposed features for recognition of speech under stress. IEEE Trans. Speech Audio Process. 8(4), 429–442 (2000)
Nwe, T., Foo, S., De Silva, L.: Speech emotion recognition using hidden Markov models. Speech Commun. 41, 603–623 (2003)
Teager, H.: Some observations on oral air flow during phonation. IEEE Trans. Acoust. Speech Sig. Process. 28(5), 599–601 (1990)
Zhou, G., Hansen, J., Kaiser, J.: Nonlinear feature based classification of speech under stress. IEEE Trans. Speech Audio Process. 9(3), 201–216 (2001)
Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Trans. Pattern Anal. Mach. Intell. 31, 39–58 (2009)
Sayedelahl, A., Araujo, P., Kamel, M.S.: Audio-visual feature-decision level fusion for spontaneous emotion estimation in speech conversations. In: International Conference on Multimedia and Expo Workshops, pp. 1–6 (2013)
Metallinou, A., Lee, S., Narayanan, S.: Decision level combination of multiple modalities for recognition and analysis of emotional expression. In: Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, pp. 2462–2465 (2010)
Nicolaou, M., Gunes, H., Pantic, M.: Audio-visual classification and fusion of spontaneous affective data in likelihood space. In: International Conference on Pattern Recognition (ICPR), pp. 3695–3699 (2010)
Lin, J.C., Wu, C.H., Wei, W.L.: Error weighted semi-coupled hidden Markov model for audio-visual emotion recognition. IEEE Trans. Multimed. 14, 142–156 (2012)
Schuller, B., et al.: AVEC 2011 the first international audio/visual emotion challenge. In: Proceedings of the First International Audio/Visual Emotion Challenge and Workshop (ACII), pp. 415–424 (2011)
Schuller, B., et al.: 3D dimensional affect and depression recognition challenge. In: Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (AVEC), ACM Multimedia Conference, Orlando, FL, USA, 3–7 November 2014 (2014)
http://cs.anu.edu.au/few/emotiw2015.html. Accessed Dec 2015
http://sspnet.eu/fera2015/. Accessed Dec 2015
McKeown, G., Valstar, M., Cowie, R., Pantic, M.: The SEMAINE corpus of emotionally coloured character interactions. In: IEEE Conference on Multimedia and Expo (2010). http://www.semaine-db.eu/
Yang, Y.-H., Lin, Y.-C., Su, Y.-F., Chen, H.-H.: A regression approach to music emotion recognition. IEEE Trans. Audio Speech Lang. Process. 16, 448–457 (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this paper
Cite this paper
Hamdy, S. (2018). Human Emotion Interpreter. In: Bi, Y., Kapoor, S., Bhatia, R. (eds) Proceedings of SAI Intelligent Systems Conference (IntelliSys) 2016. IntelliSys 2016. Lecture Notes in Networks and Systems, vol 15. Springer, Cham. https://doi.org/10.1007/978-3-319-56994-9_63
Download citation
DOI: https://doi.org/10.1007/978-3-319-56994-9_63
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-56993-2
Online ISBN: 978-3-319-56994-9
eBook Packages: EngineeringEngineering (R0)