skip to main content
research-article

Looking at near-duplicate videos from a human-centric perspective

Published:27 August 2010Publication History
Skip Abstract Section

Abstract

Popular content in video sharing websites (e.g., YouTube) is usually replicated via identical copies or near-duplicates. These duplicates are usually studied because they pose a threat to site owners in terms of wasted disk space, or privacy infringements. Furthermore, this content might potentially hinder the users' experience in these websites. The research presented in this article focuses around the central argument that there is no agreement on the technical definition of what these near-duplicates are, and, more importantly, there is no strong evidence that users of video sharing websites would like this content to be removed. Most scholars define near-duplicate video clips (NDVC) by means of non-semantic features (e.g., different image/audio quality), while a few also include semantic features (i.e., different videos of similar content). However, it is unclear what features contribute to the human perception of near-duplicate videos. The findings of four large scale online surveys that were carried out in the context of our research confirm the relevance of both types of features. Some of our findings confirm the adopted definitions of NDVC whereas other findings are surprising: Near-duplicate videos with different image quality, audio quality, or with/without overlays were perceived as NDVC. However, the same could not be verified when videos differed by more than one of these features at the same time. With respect to semantics, it is yet unclear the exact role that it plays in relation to the features that make videos alike. From a user's perspective, participants preferred in most cases to see only one of the NDVC in the search results of a video search query and they were more tolerant to changes in the audio than in the video tracks. Based on all these findings, we propose a new user-centric NDVC definition and present implications for how duplicate content should be dealt with by video sharing Web sites.

Skip Supplemental Material Section

Supplemental Material

References

  1. Basharat, A., Zhai, Y., and Shan, M. 2008. Content based video matching using spatiotemporal volumes,. J. Comput. Vis. Image Under. 110, 3, 360--377. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Benevenuto, F., Duarte, F., Rodrigues, T., Almeida, V. A., Almeida, J. M., and Ross, K. W. 2008. Understanding video interactions in youtube. In Proceeding of the 16th ACM International Conference on Multimedia (MM'08). ACM, New York, 761--764. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Bruce, B., Green, P. R., and Georgeson, M. A. 1996. Visual Perception. 3rd Ed. Psychology Press.Google ScholarGoogle Scholar
  4. Celebi, M. E. and Aslandogan, Y. A. 2005. Human perception-driven, similarity-based access to image databases. In Proceedings of the 18th International Florida Artificial Intelligence Research Society Conference. I. Russell and Z. Markov, Eds. 245--251.Google ScholarGoogle Scholar
  5. Cha, M., Kwak, H., Rodriguez, P., Ahn, Y.-Y., and Moon, S. 2007. I tube, you tube, everybody tubes: analyzing the world's largest user generated content video system. In Proceedings of the 7th ACM SIGCOMM Conference on Internet Measurement (IMC'07). ACM, New York, 1--14. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Cheng, R., Huang, Z., Shen, H. T., and Zhou, X. 2009. Interactive near-duplicate video retrieval and detection. In Proceedings of the 17th ACM International Conference on Multimedia (MM'09). ACM, New York, 1001--1002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Cheng, X. and Chia, L.-T. 2010. Stratification-based keyframe cliques for removal of near-duplicates in video search results. In Proceedings of the International Conference on Multimedia Information Retrieval (MIR'10). ACM, New York, 313--322. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Gill, P., Li, Z., Arlitt, M., and Mahanti, A. 2008. Characterizing users sessions on youtube. In Proceedings of the SPIE/ACM Conference on Multimedia Computing and Networking (MMCN).Google ScholarGoogle Scholar
  9. Guyader, N., Borgne, H. L., Hérault, J., and Guérin-Dugué, A. 2002. Towards the introduction of human perception in a natural scene classification system. In Proceedings of Neural Networks for Signal Processing. 385--394.Google ScholarGoogle Scholar
  10. Halvey, M. J. and Keane, M. T. 2007. Exploring social dynamics in online media sharing. In Proceedings of the 16th International Conference on World Wide Web (WWW'07). ACM, New York, 1273--1274. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Hsu, W. H., Kennedy, L. S., and Chang, S.-F. 2006. Video search reranking via information bottleneck principle. In Proceedings of the 14th Annual ACM International Conference on Multimedia (MULTIMEDIA'06). ACM, New York, 35--44. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Kruitbosch, G. and Nack, F. 2008. Broadcast yourself on youtube: really? In Proceeding of the 3rd ACM International Workshop on Human-Centered Computing (HCC'08). ACM, New York, 7--10. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Maia, M., Almeida, J., and Almeida, V. 2008. Identifying user behavior in online social networks. In Proceedings of the 1st Workshop on Social Network Systems (SocialNets'08). ACM, New York, 1--6. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Payne, J. S. and Stonham, T. J. 2001. Can texture and image content retrieval methods match humanperception? In Proceedings of Intelligent Multimedia, Video and Speech Processing. 154--157.Google ScholarGoogle Scholar
  15. Rui, Y., Huang, T., and Chang, S. 1999. Image retrieval: current techniques, promising directions and open issues. J. Vis. Comm. Image Repres. 10, 4, 39--62.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. seok Min, H., Choi, J., Neve, W. D., and Ro, Y. M. 2009. Near-duplicate video detection using temporal patterns of semantic concepts. In Proceedings of the International Symposium on Multimedia, 65--71. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Shao, J., Shen, H. T., and Zhou, X. 2008. Challenges and techniques for effective and efficient similarity search in large video databases. Proc. VLDB Endow. 1, 2, 1598--1603. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Shen, H. T., Zhou, X., Huang, Z., Shao, J., and Zhou, X. 2007. Uqlips: a real-time near-duplicate video clip detection system. In Proceedings of the 33rd International Conference on Very Large Data Bases (VLDB'07). VLDB Endowment, 1374--1377. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Kim, H.-S., Chang, H.-W., Lee, J., and Lee, D. 2010. Effective near-duplicate image detection using gene sequence alignment. In Advances in Information Retrieval, Lecture Notes in Computer Science, vol. 5993. Springer, Berlin, 229--240. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Tversky, A. 1977. Features of similarity. Psych. Rev. 84, 4, 327--352.Google ScholarGoogle ScholarCross RefCross Ref
  21. Wu, X., Hauptmann, A. G., and Ngo, C.-W. 2007. Practical elimination of near-duplicates from web video search. In Proceedings of the 15th International Conference on Multimedia (MULTIMEDIA'07). ACM, New York, 218--227. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Yang, X., Zhu, Q., and Cheng, K.-T. 2009. Near-duplicate detection for images and videos. In Proceedings of the 1st ACM Workshop on Large-Scale Multimedia Retrieval and Mining (LS-MMRM'09). ACM, New York, 73--80. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Zhou, X., Zhou, X., Chen, L., Bouguettaya, A., Xiao, N., and Taylor, J. A. 2009. An efficient near-duplicate video shot detection method using shot-based interest points. Trans. Multimedia. 11, 5, 879--891. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Looking at near-duplicate videos from a human-centric perspective

          Recommendations

          Reviews

          Sebastien Lefevre

          Video repositories on the Web (such as YouTube) are very popular, as they enable users to gain access to a lot of video content and also propose new content. However, this open way of feeding a database introduces a major problem: when providing new content to the repository, a given user may not have checked (assuming this was even possible) whether the proposed content was already in the database. Thus, video content is very often replicated in the video repository, which leads to several drawbacks: a waste of disk space by keeping the same data several times in the database, a bad search experience for the user (with duplicate results that are often useless), and, finally, a source of privacy-related problems. While it would be very easy to remove exact copies of the same video from a Web repository, the problem is much more difficult when dealing with very similar (but not identical) copies of a video. This paper focuses on the case of such near-duplicate videos in Web repositories. Contrary to most existing studies, the authors' proposal is not of a technical nature-that is, they do not introduce a new duplicate removal method based on image and signal processing; rather, it takes a human-centric approach. Such works, which are definitely complementary to existing works and very helpful to the computer science community, are much too rare in the literature. Here, the authors study the role of near-duplicate videos from a user's point of view and try to answer several questions related to the nature and the role of such videos. To do so, they perform a psychophysical experiment with more than 1,300 participants, in order to verify the following hypotheses related to near-duplicate videos: (1) "Video search is the main method for reaching content on video sharing Web sites." (2) "Identical or approximately identical videos ... are considered by the users as similar clips." (3) Duplicate content should not be removed from the search results. This experiment led the authors to provide a user-centric definition of near-duplicate video content: this content includes either approximately identical videos that might differ only in one or a few features (such as in encoding parameters, photometric variations, editing operations, or audio overlays), or different videos that share a visual similarity and a semantic relatedness. Even though the proposed study is too short-the experiment doesn't include enough videos-and the presentation of the results is rather flat, the paper offers an original look at the field of content-based video indexing and retrieval. Thus, I recommend this paper to computer scientists who are interested in the design of such systems. Online Computing Reviews Service

          Access critical reviews of Computing literature here

          Become a reviewer for Computing Reviews.

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in

          Full Access

          • Published in

            cover image ACM Transactions on Multimedia Computing, Communications, and Applications
            ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 6, Issue 3
            August 2010
            203 pages
            ISSN:1551-6857
            EISSN:1551-6865
            DOI:10.1145/1823746
            Issue’s Table of Contents

            Copyright © 2010 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 27 August 2010
            • Accepted: 1 June 2010
            • Revised: 1 May 2010
            • Received: 1 March 2010
            Published in tomm Volume 6, Issue 3

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article
            • Research
            • Refereed

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader