Abstract
This paper presents an example of a system for the analysis of emotions contained within radio broadcasts. We prepared training data, did feature extraction, built classifiers for music/speech discrimination and for emotion detection in music. To study changes in emotions, we used recorded broadcasts from 4 selected European radio stations. The collected data allowed us to determine the dominant emotion in the radio broadcasts and construct maps visualizing the distribution of emotions in time. The obtained results provide a new interesting view of the emotional content of radio station broadcasts.
Keywords
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Li, T., Ogihara, M.: Detecting emotion in music. In: Proceedings of the Fifth International Symposium on Music Information Retrieval, pp. 239–240 (2003)
Grekow, J., Raś, Z.W.: Detecting emotions in classical music from MIDI files. In: Rauch, J., Raś, Z.W., Berka, P., Elomaa, T. (eds.) ISMIS 2009. LNCS (LNAI), vol. 5722, pp. 261–270. Springer, Heidelberg (2009)
Lu, L., Liu, D., Zhang, H.J.: Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio, Speech and Language Processing 14(1), 5–18 (2006)
Song, Y., Dixon, S., Pearce, M.: Evaluation of Musical Features for Emotion Classification. In: Proceedings of the 13th International Society for Music Information Retrieval Conference (2012)
Yang, Y.-H., Lin, Y.C., Su, Y.F., Chen, H.H.: A regression approach to music emotion recognition. IEEE Transactions on Audio, Speech, and Language Processing 16(2), 448–457 (2008)
Schmidt, E., Kim, Y.: Modeling Musical Emotion Dynamics with Conditional Random Fields. In: Proceedings of the 12th International Society for Music Information Retrieval Conference, pp. 777–782 (2011)
Schmidt, E.M., Turnbull, D., Kim, Y.E.: Feature Selection for Content-Based, Time-Varying Musical Emotion Regression. In: Proc. ACM SIGMM International Conference on Multimedia Information Retrieval, Philadelphia, PA (2010)
Schmidt, E.M., Kim, Y.E.: Prediction of time-varying musical mood distributions from audio. In: Proceedings of the 2010 International Society for Music Information Retrieval Conference, Utrecht, Netherlands (2010)
Grekow, J.: Mood tracking of musical compositions. In: Chen, L., Felfernig, A., Liu, J., Raś, Z.W. (eds.) ISMIS 2012. LNCS (LNAI), vol. 7661, pp. 228–233. Springer, Heidelberg (2012)
Grekow, J., Raś, Z.W.: Emotion based MIDI files retrieval system. In: Raś, Z.W., Wieczorkowska, A.A. (eds.) Advances in Music Information Retrieval. SCI, vol. 274, pp. 261–284. Springer, Heidelberg (2010)
Bachorik, J.P., Bangert, M., Loui, P., Larke, K., Berger, J., Rowe, R., Schlaug, G.: Emotion in motion: Investigating the time-course of emotional judgments of musical stimuli. Music Perception 26(4), 355–364 (2009)
Xiao, Z., Dellandrea, E., Dou, W., Chen, L.: What is the best segment duration for music mood analysis? In: International Workshop on Content-Based Multimedia Indexing (CBMI 2008), pp. 17–24 (2008)
Schmidt, E.M., Scott, J.J., Kim, Y.E.: Feature Learning in Dynamic Environments: Modeling the Acoustic Structure of Musical Emotion. In: Proceedings of the 12th International Society for Music Information Retrieval Conference, pp. 325–330 (2012)
Yang, Y.H., Homer, H., Chen, H.H.: Machine Recognition of Music Emotion: A Review. ACM Transactions on Intelligent Systems and Technology 3(6), Article No. 40 (2012)
Kim, Y., Schmidt, E., Migneco, R., Morton, B., Richardson, P., Scott, J., Speck, J., Turnbull, D.: State of the Art Report: Music Emotion Recognition: A State of the Art Review. In: Proceedings of the 11th International Society for Music Information Retrieval Conference, pp. 255–266 (2010)
Mohammad, S.: From Once Upon a Time to Happily Ever After: Tracking Emotions in Novels and Fairy Tales. In: Proceedings of the ACL 2011 Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities, Portland, OR, USA, pp. 105–114 (2011)
Yeh, J.-H., Pao, T.-L., Pai, C.-Y., Cheng, Y.-M.: Tracking and Visualizing the Changes of Mandarin Emotional Expression. In: Huang, D.-S., Wunsch II, D.C., Levine, D.S., Jo, K.-H. (eds.) ICIC 2008. LNCS, vol. 5226, pp. 978–984. Springer, Heidelberg (2008)
Lidy, T., Rauber, A.: Visually Profiling Radio Stations. In: Proceedings of the 7th International Conference on Music Information Retrieval (2006)
Thayer, R.E.: The biopsychology arousal. Oxford University Press (1989)
Tzanetakis, G., Cook, P.: Marsyas: A framework for audio analysis. Organized Sound 10, 293–302 (2000)
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA Data Mining Software: An Update. SIGKDD Explorations, 11(1) (2009)
Witten, I.H., Frank, E.: Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, San Francisco (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Grekow, J. (2014). Mood Tracking of Radio Station Broadcasts. In: Andreasen, T., Christiansen, H., Cubero, JC., Raś, Z.W. (eds) Foundations of Intelligent Systems. ISMIS 2014. Lecture Notes in Computer Science(), vol 8502. Springer, Cham. https://doi.org/10.1007/978-3-319-08326-1_19
Download citation
DOI: https://doi.org/10.1007/978-3-319-08326-1_19
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-08325-4
Online ISBN: 978-3-319-08326-1
eBook Packages: Computer ScienceComputer Science (R0)