Skip to main content

Population Codes Representing Musical Timbre for High-Level fMRI Categorization of Music Genres

  • Conference paper
Machine Learning and Interpretation in Neuroimaging

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 7263))

Abstract

We present experimental evidence in support of distributed neural codes for timbre that are implicated in discrimination of musical styles. We used functional magnetic resonance imaging (fMRI) in humans and multivariate pattern analysis (MVPA) to identify activation patterns that encode the perception of rich music audio stimuli from five different musical styles. We show that musical styles can be automatically classified from population codes in bilateral superior temporal sulcus (STS). To investigate the possible link between the acoustic features of the auditory stimuli and neural population codes in STS, we conducted a representational similarity analysis and a multivariate regression-retrieval task. We found that the similarity structure of timbral features of our stimuli resembled the similarity structure of the STS more than any other type of acoustic feature. We also found that a regression model trained on timbral features outperformed models trained on other types of audio features. Our results show that human brain responses to complex, natural music can be differentiated by timbral audio features, emphasizing the importance of timbre in auditory perception.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 49.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Norman, K., Polyn, S.M., Detre, G.J., Haxby, J.V.: Beyond mind-reading: multi-voxel pattern analysis of fMRI data. Trends in Cognitive Sciences 10(9), 424–430 (2006)

    Article  Google Scholar 

  2. Staeren, N., Renvall, H., De Martino, F., Goebel, R., Formisano, E.: Sound categories are represented as distributed patterns in the human auditory cortex. Current Biology 19(6), 498–502 (2009)

    Article  Google Scholar 

  3. Kilian-Hütten, N., Valente, G., Vroomen, J., Formisano, E.: Auditory cortex encodes the perceptual interpretation of ambiguous sound. The Journal of Neuroscience 31(5), 1715–1720 (2011)

    Article  Google Scholar 

  4. Lee, Y.-S., Janata, P., Frost, C., Hanke, M., Granger, R.: Investigation of melodic contour processing in the brain using multivariate pattern-based fMRI. NeuroImage 57(1), 293–300 (2011)

    Article  Google Scholar 

  5. Samson, S., Zatorre, R.J.: Melodic and harmonic discrimination following unilateral cerebral excision. Brain and Cognition 7(3), 348–360 (1988)

    Article  Google Scholar 

  6. Bizley, J.K., Walker, K.M.M.: Sensitivity and selectivity of neurons in auditory cortex to the pitch, timbre, and location of sounds. The Neuroscientist 16(4), 453–469 (2010)

    Article  Google Scholar 

  7. Samson, S., Zatorre, R.J.: Contribution of the right temporal lobe to musical timbre discrimination. Neuropsychologia 32(2), 231–240 (1994)

    Article  Google Scholar 

  8. Warren, J.D., Jennings, A.R., Griffiths, T.D.: Analysis of the spectral envelope of sounds by the human brain. NeuroImage 24(4), 1052–1057 (2005)

    Article  Google Scholar 

  9. Meyer, M., Zysset, S., von Cramon, D.Y., Alter, K.: Distinct fMRI responses to laughter, speech, and sounds along the human peri-sylvian cortex. Cognitive Brain Research 24(2), 291–306 (2005)

    Article  Google Scholar 

  10. Buracas, G.T., Boynton, G.M.: Efficient design of event-related fMRI experiments using m-sequences. NeuroImage 16(3), 801–813 (2002)

    Article  Google Scholar 

  11. Cox, R.W.: AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. Computers and Biomedical Research 29(3), 162–173 (1996)

    Article  Google Scholar 

  12. Menon, V.: Neural Correlates of Timbre Change in Harmonic Sounds. NeuroImage 17(4), 1742–1754 (2002)

    Article  Google Scholar 

  13. Hanson, S.J., Matsuka, T., Haxby, J.V.: Combinatorial codes in ventral temporal lobe for object recognition: Haxby (2001) revisited: Is there a face area? Neuroimage (23), 156–166 (2004)

    Article  Google Scholar 

  14. Kriegeskorte, N., Mur, M., Ruff, D., Kiani, P., Bodurka, J., Esteky, H.: Matching categorical object representations in inferior temporal cortex of man and monkey. Neuron (60), 1126–1141 (2008)

    Article  Google Scholar 

  15. Casey, M.A.: Bregman music and auditory python toolbox (January 2012), http://bregman.dartmouth.edu/bregman

  16. Müller, M., Ewert, S., Kreuzer, S.: Making chroma features more robust to timbre changes. In: Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, pp. 1877–1880. IEEE (2009)

    Google Scholar 

  17. Brown, J.C., Puckette, M.S.: An efficient algorithm for the calculation of a constant q transform. Journal of the Acoustical Society of America (92), 2698–2701 (1992)

    Article  Google Scholar 

  18. Logan, B.: Mel frequency cepstral coefficients for music modeling. In: Proceedings of the International Symposium on Music Information Retrieval (2000)

    Google Scholar 

  19. Mitchell, T.M., Shinkareva, S.V., Carlson, A., Chang, K.M., Malave, V.L., Mason, R.A., Just, M.A.: Predicting human brain activity associated with the meanings of nouns. Science 320(5880), 1191–1195 (2008)

    Article  Google Scholar 

  20. Tzanetakis, G., Cook, P.R.: Musical genre classification of audio signals. IEEE Transactions on Speech and Audio Processing 10(5), 293–302 (2002)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Casey, M., Thompson, J., Kang, O., Raizada, R., Wheatley, T. (2012). Population Codes Representing Musical Timbre for High-Level fMRI Categorization of Music Genres. In: Langs, G., Rish, I., Grosse-Wentrup, M., Murphy, B. (eds) Machine Learning and Interpretation in Neuroimaging. Lecture Notes in Computer Science(), vol 7263. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-34713-9_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-34713-9_5

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-34712-2

  • Online ISBN: 978-3-642-34713-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics