Abstract
This paper concerns the problem of the effect of emotional change on humans and machines for speaker identification. A contrasting experiment is carried out between Automatic Speaker Identification (ASI) system (applying GMM-UBM and Emotional Factor Analysis (EFA) algorithm) and aural system on emotional speech corpus MASC. The experimental result is similar to that in channel-mismatched condition, i.e. the ASI system is much better than the single listener, especially when emotion compensation algorithm EFA is applied. Meanwhile,fusion of multiple listeners can significantly improve the aural system performance by 23.86% and make it outperform the ASI system.
This paper is supported by NSFC60970080 and the Special Funds for Key Program of the China No. 2009ZX01039-002-001-04.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Schmidt-Nielsen, A., Crystal, T.H.: Speaker verification by human listeners: experiments comparing human and machine performance using th2e NIST 1998 speaker verification data. Digital Signal Processing 10, 249–266 (2000)
Kajarekar, S.S., Bratt, H., Shriberg, E., de Leon, R.: A study of intentional voice modifications for evading automatic speaker recognition. In: Speaker Odessy (2006)
Hautamaki, V., Kinnunen, T., Nosratighods, M., Lee, K.-A., Ma, B., Li, H.: Approaching human listener accuracy with modern speaker verification. In: Interspeech 2010, pp. 1473–1476 (2010)
The NIST year, speaker recognition evaluation plan (2010)
Shriberg, E., Graciarena, M., Bratt, H., Kathol, A., Kajarekar, S., Jameel, H., Richey, C., Goodman, F.: Effects of vocal effort and speaking style on text-independent speaker verification. In: Interspeech 2007, Antwerp, pp. 950–954 (2007)
Wu, T., Yang, Y., Wu, Z., Li, D.: MASC: A Speech Corpus in Mandarin for Emotion Analysis and Affective Speaker Recognition. In: ODYSSEY 2006, pp. 1–5 (June 2006)
Kenny, P., Boulianne, G., Ouellet, P., Dumouchel, P.: Joint factor analysis versus eigenchannels in speaker recognition. IEEE Transaction on Audio Speech and Language Processing 15(4), 1435–1447 (2007)
Chen, L., Yang, Y.: Applying Emotional Factor Analysis and I-Vector to Emotional Speaker Recognition. Submitted to CCBR (2011)
http://speech.fit.vutbr.cz/en/software/joint-factor-analysis-matlab-demo
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Yang, Y., Chen, L., Wang, W. (2011). Emotional Speaker Identification by Humans and Machines. In: Sun, Z., Lai, J., Chen, X., Tan, T. (eds) Biometric Recognition. CCBR 2011. Lecture Notes in Computer Science, vol 7098. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-25449-9_21
Download citation
DOI: https://doi.org/10.1007/978-3-642-25449-9_21
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-25448-2
Online ISBN: 978-3-642-25449-9
eBook Packages: Computer ScienceComputer Science (R0)