Skip to main content

Piano and Guitar Tone Distinction Based on Extended Feature Analysis

  • Conference paper
  • First Online:
Classification and Data Mining

Abstract

In this work single piano and guitar tones are distinguished by means of various features of the music time series. In a first study, three different kinds of high-level features and MFCC are taken into account to classify the piano and guitar tones. The features are called high-level because they try to reflect the physical structure of a musical instrument on temporal and spectral levels. In our study, three spectral features and one temporal feature are used for the classification task. The spectral features characterize the distribution of overtones, the temporal feature the energy of a tone. In a second study as many low level and the high level features as possible proposed in the literature are combined for the classification task.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Akaike, H. (1974). A new look at the statistical model identification. IEEE Transactions on Automatic Control, 19(6), 716–723. doi:10.1109/TAC.1974.1100705.

    Article  MathSciNet  MATH  Google Scholar 

  • Bischl, B. (2011). Machine learning in R, R-package, TU Dortmund. http://r-forge.r-project.org/projects/mlr/.

  • Bischl, B., Vatolkin, I., & Preuss, M. (2010). Selecting small audio feature sets in music classification by means of asymmetric mutation. In: Proceedings of the 11th International Conference on Parallel Problem Solving From Nature (PPSN), Krakow, pp. 314–323.

    Google Scholar 

  • Brown, J. C., Houix, O., & McAdams, S. (2001). Feature dependence in the automatic identification of musical woodwind instruments. Journal of the Acoustical Society of America, 109, 1064–1072.

    Article  Google Scholar 

  • Fletcher, N. H. (2008). The physics of musical instruments. New York: Springer.

    Google Scholar 

  • Goto, M., Hashiguchi, H., Nishimura, T., & Oka, R. (2003). RWC music database: Music genre database and musical instrument sound database. In: ISMIR 2003 Proceedings, Baltimore, pp. 229–230.

    Google Scholar 

  • Krey, S., & Ligges, U. (2010). SVM based instrument and timbre classification. In Locarek-Junge, H., & Weihs, C. (Eds.), Classification as a tool for research. Berlin/Heidelberg/New York: Springer.

    Google Scholar 

  • Lartillot, O., & Toiviainen, P. (2007). A matlab toolbox for musical feature extraction from audio. In: International Conference on Digital Audio Effects, Bordeaux.

    Google Scholar 

  • Ligges, U. (2010). tuneR–analysis of music. http://r-forge.r-project.org/projects/tuner.

  • Livshin, A., & Rodet, X. (2006). The significance of the non-harmonic “Noise” versus the harmonic series for musical instrument recognition. In: ISMIR 2006 Proceedings, Victoria, pp. 95–100.

    Google Scholar 

  • Makhoul, J. (1975). Linear prediction: A tutorial review. IEEE, 63, 56.

    Article  Google Scholar 

  • McGill University. (2010). McGill master samples collection on DVD. http://www.music.mcgill.ca/resources/mums/html.

  • Rabiner, L., & Juang B. H. (1993). Fundamentals of speech recognition. Englewood Cliffs: Prentice Hall PTR.

    Google Scholar 

  • Theimer, W., Vatolkin, I., & Eronen, A. (2008). Definitions of audio features for music content description (Tech. Rep. TR08-2-001) University of Dortmund, Chair of Algorithm Engineering.

    Google Scholar 

  • University of Iowa. (2010). Electronic music studios. Musical instrument samples. http://theremin.music.uiowa.edu.

  • Vatolkin, I., Theimer, W., & Botteck, M. (2010). AMUSE (Advanced MUSic Explorer)–A multitool framework for music data analysis. In Proceedings of the 11th International Society for Music Information Retrieval Conference (ISMIR), Utrecht, pp. 33–38.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Markus Eichhoff .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Eichhoff, M., Vatolkin, I., Weihs, C. (2013). Piano and Guitar Tone Distinction Based on Extended Feature Analysis. In: Giusti, A., Ritter, G., Vichi, M. (eds) Classification and Data Mining. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-28894-4_26

Download citation

Publish with us

Policies and ethics