Skip to main content

Music Genre Prediction by Low-Level and High-Level Characteristics

  • Conference paper
  • First Online:
Data Analysis, Machine Learning and Knowledge Discovery

Abstract

For music genre prediction typically low-level audio signal features from time, spectral or cepstral domains are taken into account. Another way is to use community-based statistics such as Last.FM tags. Whereas the first feature group often can not be clearly interpreted by listeners, the second one lacks in erroneous or not available data for less popular songs. We propose a two-level approach combining the specific advantages of the both groups: at first we create high-level descriptors which describe instrumental and harmonic characteristics of music content, some of them derived from low-level features by supervised classification or from analysis of extended chroma and chord features. The experiments show that each categorization task requires its own feature set.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Ahrendt, P. (2006). Music Genre Classification Systems—A Computational Approach (Ph.D. thesis). Informatics and Mathematical Modelling, Technical University of Denmark, Kongens Lyngby.

    Google Scholar 

  • Berenzweig, A., Ellis, D. P. W., & Lawrence, S. (2003). Anchor space for classification and similarity measurement of music. In Proceedings of the 2003 International Conference on Multimedia and Expo (ICME) (pp. 29–32). New York: IEEE.

    Google Scholar 

  • Bertin-Mahieux, T., Eck, D., Maillet, F., & Lamere, P. (2008). Autotagger: A model for predicting social tags from acoustic features on large music databases. Journal of New Music Research, 37(2), 115–135.

    Article  Google Scholar 

  • Lartillot, O., & Toivainen, P. (2007). MIR in Matlab (ii): A toolbox for musical feature extraction from audio. In Proceedings of the 8th International Conference on Music Information Retrieval (ISMIR) (pp. 127–130). Austrian Computer Society (OCG).

    Google Scholar 

  • Mauch, M., & Dixon, S. (2010). Approximate note transcription for the improved identification of difficult chords. In Proceedings of the 11th International Society for Music Information Retrieval Conference (ISMIR) (pp. 135–140). International Society for Music Information Retrieval.

    Google Scholar 

  • Mckay, C., & Fujinaga, I. (2008). Combining features extracted from audio, symbolic and cultural sources. In Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR) (pp. 597–602). Philadelphia: Drexel University.

    Google Scholar 

  • Müller, M., & Ewert, S. (2010). Towards timbre-invariant audio features for harmony-based music. IEEE Transactions on Audio, Speech, and Language Processing, 18(3), 649–662.

    Article  Google Scholar 

  • Rötter, G., Vatolkin, I., & Weihs, C. (2011). Computational prediction of high-level descriptors of music personal categories. In Proceedings of the 2011 GfKl 35th Annual Conference of the German Classification Society (GfKl). Berlin: Springer.

    Google Scholar 

  • Theimer, W., Vatolkin, I., & Eronen, A. (2008). Definitions of audio features for music content description. Algorithm Engineering Report TR08-2-001, Technische Universität Dortmund.

    Google Scholar 

  • Vatolkin, I., Preuß, M., & Rudolph, G. (2011). Multi-objective feature selection in music genre and style recognition tasks. In Proceedings of the 2011 Genetic and Evolutionary Computation Conference (GECCO) (pp. 411–418). New York: ACM.

    Google Scholar 

  • Vatolkin, I., Preuß, M., Rudolph, G., Eichhoff, M., & Weihs, C. (2012). Multi-objective evolutionary feature selection for instrument recognition in polyphonic audio mixtures. Soft Computing, 16(12), 2027–2047.

    Article  Google Scholar 

  • Vatolkin, I., Theimer, M., & Botteck, M. (2010). AMUSE (Advanced MUSic Explorer): A multitool framework for music data analysis. In Proceedings of the 11th International Society on Music Information Retrieval Conference (ISMIR) (pp. 33–38). International Society for Music Information Retrieval.

    Google Scholar 

  • Weihs, C., Ligges, F., Mörchen, F., & Müllensiefen, D. (2007). Classification in music research. Advances in Data Analysis and Classification, 1(3), 255–291.

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

We thank the Klaus Tschira Foundation for the financial support.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Igor Vatolkin .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Vatolkin, I., Rötter, G., Weihs, C. (2014). Music Genre Prediction by Low-Level and High-Level Characteristics. In: Spiliopoulou, M., Schmidt-Thieme, L., Janning, R. (eds) Data Analysis, Machine Learning and Knowledge Discovery. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Cham. https://doi.org/10.1007/978-3-319-01595-8_46

Download citation

Publish with us

Policies and ethics