skip to main content
10.1145/2072529.2072533acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

A tempo-sensitive music search engine with multimodal inputs

Published: 30 November 2011 Publication History

Abstract

This paper presents TMSE: a novel Tempo-sensitive Music Search Engine with multimodal inputs for wellness and therapeutic applications. TMSE integrates six different interaction modes, Query-by-Number, Query-by-Sliding, Query-by-Example, Query-by-Tapping, Query-by-Clapping, and Query-by-Walking, into one single interface for narrowing the intention gap when a user searches for music by tempo. Our preliminary evaluation results indicate that multimodal inputs of TMSE enable users to formulate tempo related queries more easily in comparison with existing music search engines.

References

[1]
Davies, M. E. P. and Plumbley, M. D. Context-Dependent Beat Tracking of Musical Audio, Audio, Speech and Language Processing, IEEE Transactions on {see also Speech and Audio Processing, IEEE Transactions on} (15:3), 2007, pp. 1009--1020
[2]
Dixon, S. Evaluation of the Audio Beat Tracking System BeatRoot, Journal of New Music Research (36:1), 2007, pp. 39--50.
[3]
Dixon, S. Onset detection revisited, Proceedings of the 9th International Conference on Digital Audio Effects, 2006, pp. 133--137.
[4]
Dixon, S. Automatic Extraction of Tempo and Beat from Expressive Performances, Journal of New Music Research (30), 2001, pp. 39--58.
[5]
Drake, C., Penel, A. and Bigand, E. Tapping in time with mechanically and expressively performed music, Music Perception, 2000, pp. 1--23.
[6]
Ellis, D. P. W. Beat Tracking by Dynamic Programming, Journal of New Music Research (36:1), 2007, pp. 51--60.
[7]
Klapuri, A. P., Eronen, A. J. and Astola, J. T. Analysis of the meter of acoustic musical signals," Audio, Speech and Language Processing, IEEE Transactions on {see also Speech and Audio Processing, IEEE Transactions on} (14:1), 2006, pp. 342--355
[8]
Thaut, M. H., Mcintosh, G. C., Rice, R. R., Miller, R. A., Rathbun, J. and Brault, J. M. Rhythmic auditory stimulation in gait training for Parkinson's disease patients, Movement Disorders (11:2), 1996, pp. 193--200.
[9]
Hanna, P., & Robine, M. (2009). Query by tapping system based on alignment algorithm. ICASSP 2009, pp. 1881--1884).
[10]
Jang, J.S., Lee, H.R., & Yeh, C.-H. (2001). Query by tapping: A new paradigm for Content-Based music retrieval from acoustic input. In H.-Y. Shum, M. Liao, & S.-F. Chang (Eds.) Advances in Multimedia Information Processing, PCM 2001, vol. 2195 of Lecture Notes in Computer Science, chap. 76, (pp. 590--597). (
[11]
Wang, A. (2006). The Shazam music recognition service. Commun. ACM, 49(8), 44--48. (URL http://dx.doi.org/10.1145/1145287.1145312.
[12]
The midomi music search: http://www.midomi.com
[13]
R. Typke, F. Wiering, and R. Veltkamp, "A survey of music information retrieval systems," ISMIR, 2005, pp. 153--160.
[14]
Oliver, N. and Flores-Mangas, F. MPTrain: a mobile, music and physiology-based personal trainer. Proceedings of the 8th conference on Human-computer interaction with mobile devices and services, ACM, 2006, pp. 21--28.
[15]
Li, Z., Xiang, Q., Hockman, J., Yang, J., Yi, Y., Fujinaga, I., & Wang, Y. (2010). A music search engine for therapeutic gait training. In Proceedings of the international conference on Multimedia, MM '10, (pp. 627--630). New York, NY, USA: ACM.
[16]
"WaoN: A wave-to-notes transcriber," {Web site} 2006, {2008 Nov 20}, Available: www.kichiki.com/WAON
[17]
Gaver, W. and Smith, R. Auditory icons in large-scale collaborative environments, ACM SIGCHI Bulletin (23:1), 1991, pp. 96.
[18]
Brewster, S., Wright, P. and Edwards, A. An evaluation of earcons for use in auditory human-computer interfaces. Proceedings of the INTERACT'93 and CHI'93 conference on Human factors in computing systems, ACM, 1993, pp. 222--227.
[19]
Zhao, S., Dragicevic, P., Chignell, M., Balakrishnan, R. and Baudisch, P. Earpod: eyes-free menu selection using touch input and reactive audio feedback, Proceedings of the SIGCHI conference on Human factors in computing systems, ACM, 2007, pp. 1395--1404.
[20]
Brodsky, W. (2005). The effects of metronomic pendular adjustment versus tap-tempo input on the stability and accuracy of tempo perception. Cognitive Processing, 6(2), 117--127.
[21]
X. S. Hua, M. Worring, and T. S. Chua, "Internet multimedia search and mining", Bentham Science Publishers, 2010.

Cited By

View all
  • (2014)Quantifying Auditory Temporal Stability in a Large Database of Recorded MusicPLoS ONE10.1371/journal.pone.01104529:12(e110452)Online publication date: 3-Dec-2014
  • (2013)Multi-Functional Music Auxiliary System Based on Voice Recognition on Tablet PCInternational Asia Conference on Industrial Engineering and Management Innovation (IEMI2012) Proceedings10.1007/978-3-642-38445-5_171(1629-1637)Online publication date: 29-May-2013
  • (2011)1st international ACM workshop on music information retrieval with user-centered and multimodal strategies (MIRUM)Proceedings of the 19th ACM international conference on Multimedia10.1145/2072298.2072386(603-604)Online publication date: 28-Nov-2011

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MIRUM '11: Proceedings of the 1st international ACM workshop on Music information retrieval with user-centered and multimodal strategies
November 2011
70 pages
ISBN:9781450309868
DOI:10.1145/2072529
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 30 November 2011

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. multimodal query
  2. music information retrieval
  3. tempo

Qualifiers

  • Research-article

Conference

MM '11
Sponsor:
MM '11: ACM Multimedia Conference
November 30, 2011
Arizona, Scottsdale, USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)1
Reflects downloads up to 10 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2014)Quantifying Auditory Temporal Stability in a Large Database of Recorded MusicPLoS ONE10.1371/journal.pone.01104529:12(e110452)Online publication date: 3-Dec-2014
  • (2013)Multi-Functional Music Auxiliary System Based on Voice Recognition on Tablet PCInternational Asia Conference on Industrial Engineering and Management Innovation (IEMI2012) Proceedings10.1007/978-3-642-38445-5_171(1629-1637)Online publication date: 29-May-2013
  • (2011)1st international ACM workshop on music information retrieval with user-centered and multimodal strategies (MIRUM)Proceedings of the 19th ACM international conference on Multimedia10.1145/2072298.2072386(603-604)Online publication date: 28-Nov-2011

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media