Abstract
The recognition of emotion in human speech has gained increasing attention in recent years due to the wide variety of applications that benefit from such technology. Detecting emotion from speech can be viewed as a classification task. It consists of assigning, out of a fixed set, an emotion category e.g. happiness, anger, to a speech utterance. In this paper, we have tackled two emotions namely happiness and anger. The parameters extracted from speech signal depend on speaker, spoken word as well as emotion. To detect the emotion, we have kept the spoken utterance and the speaker constant and only the emotion is changed. Different features are extracted to identify the parameters responsible for emotion. Wavelet packet transform (WPT) is found to be emotion specific. We have performed the experiments using three methods. Method uses WPT and compares the number of coefficients greater than threshold in different bands. Second method uses energy ratios of different bands using WPT and compares the energy ratios in different bands. The third method is a conventional method using MFCC. The results obtained using WPT for angry, happy and neutral mode are 85 %, 65 % and 80 % respectively as compared to results obtained using MFCC i.e. 75 %, 45 % and 60 % respectively for the three emotions. Based on WPT features a model is proposed for emotion conversion namely neutral to angry and neutral to happy emotion.
Similar content being viewed by others
References
Agarwal, A., Jain, A., Prakash, N., & Agrawal, S. S. (2010). Word based emotion conversion in Hindi language. In ICCSIT’2010 proceedings, Chengdu (pp. 419–423).
Burkhardt, F., Polzehl, T., Stegmann, J., Metze, F., & Huber, R. (2009). Detecting real life anger. In ICASSP’09 proceedings, Taipei, Taiwan (pp. 4761–4764).
Hidayati, R., Purnama, I. K. E., & Purnomo, M. H. (2009). The extraction of acoustic features of infant cry for emotion detection based on pitch and formants. In ICICI-BME’09, proceedings, Bandung (pp. 1–5).
Krajewski, J., Batliner, A., & Kessel, S. (2010). Comparing multiple classifiers for speech-based detection of self-confidence—a pilot study. In ICPR’2010 proceedings, Istanbul (pp. 3716–3719).
Laskowski, K. (2010). Finding emotionally involved speech using implicitly proximity-annotated laughter. In ICASSP’2010 proceedings, Dallas, TX (pp. 5226–5229).
Meshram, A. P., Shirbahadurkar, S. D., Kohok, A., & Jadhav, S. (2010). An overview and preparation for recognition of emotion from speech signal with multi modal fusion. In ICCAE’2010 proceedings, Singapore (pp. 446–452).
Metze, F., Polzehl, T., & Wagner, M. (2009). Fusion of acoustic and linguistic features for emotion detection. In ICSC’09 proceedings, Berkeley (pp. 153–160).
Tawari, A., & Trivedi, M. M. (2010). Speech emotion analysis: exploring the role of context. IEEE Transactions on Multimedia, 12(6), 502–509.
Wooil, K., & Hansen, J. H. L. (2010). Angry emotion detection from real-life conversational speech by leveraging content structure. In ICASSP’2010, proceedings, Dallas, TX (pp. 5166–5169).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Degaonkar, V.N., Apte, S.D. Emotion modeling from speech signal based on wavelet packet transform. Int J Speech Technol 16, 1–5 (2013). https://doi.org/10.1007/s10772-012-9142-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10772-012-9142-8