ABSTRACT
We describe a system for synchronization and organization of user-contributed content from live music events. We start with a set of short video clips taken at a single event by multiple contributors, who were using a varied set of capture devices. Using audio fingerprints, we synchronize these clips such that overlapping clips can be displayed simultaneously. Furthermore, we use the timing and link structure generated by the synchronization algorithm to improve the findability and representation of the event content, including identifying key moments of interest and descriptive text for important captured segments of the show. We also identify the preferred audio track when multiple clips overlap. We thus create a much improved representation of the event that builds on the automatic content match. Our work demonstrates important principles in the use of content analysis techniques for social media content on the Web, and applies those principles in the domain of live music capture.
- M. G. Christel, A. G. Hauptmann, and H. D. Wactlar. Collages as dynamic summaries for news video. In MULTIMEDIA '02: Proceedings of the 10th international conference on Multimedia, pages 561{569. ACM Press, 2002. Google ScholarDigital Library
- S. J. Cunningham and D. M. Nichols. How people find videos. In JCDL '08: Proceedings of the Eigth ACM/IEEE joint conference on Digital libraries, New York, NY, USA, 2008. ACM. Google ScholarDigital Library
- A. Graham, H. Garcia-Molina, A. Paepcke, and T. Winograd. Time as essence for photo browsing through personal digital libraries. In Proceedings of the Second ACM/IEEE-CS Joint Conference on Digital Libraries, 2002. Google ScholarDigital Library
- J. Haitsma and T. Kalker. A Highly Robust Audio Fingerprinting System With an Efficient Search Strategy. Journal of New Music Research, 32(2):211--221, 2003.Google ScholarCross Ref
- C. Herley. Accurate repeat finding and object skipping using fingerprints. In MULTIMEDIA '05: Proceedings of the 13th international conference on Multimedia, pages 656--665. ACM Press, 2005. Google ScholarDigital Library
- C. Herley. ARGOS: automatically extracting repeating objects from multimedia streams. IEEE Transactions on Multimedia, 8(1):115--129, 2006. Google ScholarDigital Library
- V. Kaplun, P. Vora, M. Naaman, P. Mead, and A. Moed. Understanding media capture and consumption for live music events. Technical report, Yahoo! Inc., 2008. In Submission.Google Scholar
- L. Kennedy and M. Naaman. Generating diverse and representative image search results for landmark queries. In Proceedings of the Seventeenth International World Wide Web Conference, New York, NY, USA, 2008. ACM. Google ScholarDigital Library
- M. Naaman, Y. J. Song, A. Paepcke, and H. Garcia-Molina. Automatic organization for digital photographs with geographic coordinates. In Proceedings of the Fourth ACM/IEEE-CS Joint Conference on Digital Libraries, 2004. Google ScholarDigital Library
- S. U. Naci and A. Hanjalic. Intelligent browsing of concert videos. In MULTIMEDIA '07: Proceedings of the 15th international conference on Multimedia, pages 150--151, New York, NY, USA, 2007. ACM. Google ScholarDigital Library
- R. Nair, N. Reid, and M. Davis. Photo LOI: Browsing multi-user photo collections. In Proceedings of the 13th International Conference on Multimedia (MM2005). ACM Press, 2005. Google ScholarDigital Library
- J. Ogle and D. Ellis. Fingerprinting to Identify Repeated Sound Events in Long-Duration Personal Audio Recordings. Acoustics, Speech and Signal Processing, 2007.Google Scholar
- ICASSP 2007. IEEE International Conference on, 1, 2007.Google Scholar
- H. W. Prarthana Shrestha, Mauro Barbieri. Synchronization of multi-camera video recordings based on audio. In MULTIMEDIA '07: Proceedings of the 15th international conference on Multimedia, pages 545--548. ACM Press, 2007. Google ScholarDigital Library
- D. Shamma, R. Shaw, P. Shafton, and Y. Liu. Watch what I watch: using community activity to understand content. Proceedings of the international workshop on Workshop on multimedia information retrieval, pages 275--284, 2007. Google ScholarDigital Library
- R. Shaw and P. Schmitz. Community annotation and remix: a research platform and pilot deployment. In HCM '06: Proceedings of the 1st ACM international workshop on Human-centered multimedia, pages 89--98, New York, NY, USA, 2006. ACM. Google ScholarDigital Library
- J. Shi and J. Malik. Normalized Cuts and Image Segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 888--905, 2000. Google ScholarDigital Library
- P. Shrestha, H. Weda, M. Barbieri, and D. Sekulovski. Synchronization of multiple video recordings based on still camera ashes. In MULTIMEDIA '06: Proceedings of the 14th international conference on Multimedia, pages 137--140. ACM Press, 2006. Google ScholarDigital Library
- C. Snoek, M. Worring, A. Smeulders, and B. Freiburg. The role of visual content and style for concert video indexing. Multimedia and Expo, 2007 IEEE International Conference on, pages 252--255, 2-5 July 2007.Google ScholarCross Ref
- S. Uchihashi, J. Foote, and A. Girgensohn. Video manga: generating semantically meaningful video summaries. In MULTIMEDIA '99: Proceedings of the 7th international conference on Multimedia, pages 383--392. ACM Press, 1999. Google ScholarDigital Library
- Y. van Houten, U. Naci, B. Freiburg, R. Eggermont, S. Schuurman, D. Hollander, J. Reitsma, M. Markslag, J. Kniest, M. Veenstra, and A. Hanjalic. The multimedian concert-video browser. Multimedia and Expo, 2005. ICME 2005. IEEE International Conference on, pages 1561--1564, 6-6 July 2005.Google ScholarCross Ref
- A. Wang. An Industrial Strength Audio Search Algorithm. In Proceedings of the International Conference on Music Information Retrieval, 2003.Google Scholar
- U. Westermann and R. Jain. Toward a common event model for multimedia applications. IEEE Multimedia, 14(1):19--29, 2007. Google ScholarDigital Library
- Youtube.com, google inc. http://www.youtube.com.Google Scholar
- A. Zunjarwad, H. Sundaram, and L. Xie. Contextual wisdom: social relations and correlations for multimedia event annotation. Proceedings of the 15th international conference on Multimedia, pages 615--624, 2007. Google ScholarDigital Library
Index Terms
- Less talk, more rock: automated organization of community-contributed collections of concert videos
Recommendations
Synchronization of multi-camera video recordings based on audio
MM '07: Proceedings of the 15th ACM international conference on MultimediaAn increasing number of people regularly capture video in social occasions like weddings, parties and holiday trips. As a result, multiple video recordings are made from a single event providing different view angles and wider coverage. This gives an ...
Blind Clustering of Music Recordings Based on Audio Fingerprinting
IIH-MSP '09: Proceedings of the 2009 Fifth International Conference on Intelligent Information Hiding and Multimedia Signal ProcessingAlthough multiple music recordings may sound identical to a human listener, the underlying representations of sound may differ due to the variations in their audio encoding and/or transmission methods. In contrast to the existing audio-fingerprinting ...
Resynchronize Japanese "Geisha" Dance Video Using Music of Different Styles
CULTURECOMPUTING '13: Proceedings of the 2013 International Conference on Culture and ComputingMusic and dancing are two different arts yet inseparable and which can be both powerful expression channels for a society or an artist. In most cases, the rhythm, tempo and performance of a dance depend on those of the music. This paper presents a new ...
Comments