Skip to main content
Log in

Video archaeology: understanding video manipulation history

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Facing the explosive growth of near-duplicate videos, video archaeology is quite desired to investigate the history of the manipulations on these videos. With the determination of derived videos according to the manipulations, a video migration map can be constructed with the pair-wise relationships in a set of near-duplicate videos. In this paper, we propose an improved video archaeology (I-VA) system by extending our previous work (Shen et al. 2010). The extensions include more comprehensive video manipulation detectors and improved techniques for these detectors. Specially, the detectors are used for two categories of manipulations, i.e., semantic-based manipulations and non-semantic-based manipulations. Moreover, the improved detecting algorithms are more stable. The key of I-VA is the construction of a video migration map, which represents the history of how near-duplicate videos have been manipulated. There are various applications based on the proposed I-VA system, such as better understanding of the meaning and context conveyed by the manipulated videos, improving current video search engines by better presentation based on the migration map, and better indexing scheme based on the annotation propagation. The system is tested on a collection of 12,790 videos and 3,481 duplicates. The experimental results show that I-VA can discover the manipulation relation among the near-duplicate videos effectively.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

References

  1. Aradilla G, Vepa J, Bourlard H (2006) Using posterior-based features in template matching for speech recognition. In: Proc. of the international conference on spoken language processing. Pittsburgh, USA, pp 2570–2573

  2. Boreczky JS, Rowe LA (1996) Comparison of video shot boundary detection techniques. In: Proc. of storage and retrieval for still image and video databases, pp 170–179

  3. Cheung SC, Zakhor A (2003) Efficient video similarity measurement with video signature. IEEE Trans Circuits Syst Video Technol 13(1):59–74

    Article  Google Scholar 

  4. Cheung SC, Zakhor A (2005) Fast similarity search and clustering of video sequences on world-wide-web. IEEE Trans Circuits Syst Video Technol 7(3):524–537

    Google Scholar 

  5. Google video. Available: http://video.google.com. Accessed Oct 2011

  6. Hampapur A, Bolle R (2002) Comparison of sequence matching techniques for video copy detection, pp 194–201

  7. Kashino K, Kurozumi T, Murase H (2003) A quick search method for audio and video signals based on histogram pruning. IEEE Trans Multimedia 5(3):348–357

    Article  Google Scholar 

  8. Kennedy L, Chang S-F (2008) Internet image archaeology: automatically tracing the manipulation history of photographs on the web. In: Proc. of the 16th ACM international conference on multimedia, pp 349–358

  9. Laptev I, Lindeberg T (2002) Space-time interest points, pp 128–142

  10. Lowe D (2003) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 20:91–110

    Google Scholar 

  11. Marziliano P, Dufaux F, Winkler S, Ebrahimi T (2002) A no-reference perceptual blur metric. In: Proc. of the international conference on image processing, vol 3, pp 57–60

  12. Meinedo H, Neto J (2003) Audio segmentation, classification and clustering in a broadcast news task. In: Proc. of the international conference on acoustics, speech, and signal processing, vol 2, pp 5–8

  13. Mikolajczyk K, Schmid C (2002) An affine invariant interest point detector. Copenhagen, Denmark, pp 128–142

  14. Mikolajczyk K, Schmid C (2003) An performance evaluation of local descriptors. In: Proc. of IEEE international conference on computer vision and pattern recognition, pp 257–264

  15. Paisitkriangkrai S, Mei T, Zhang J, Hua X-S (2010) Scalable clip-based near-duplicate video detection with ordinal measure. In: Proc. of ACM international conference on image and video retrieval. Xi’an, China, pp 121–128

  16. Sakoe H, Chiba S (1971) Recognition of continuously spoken words based on time normalization by dynamic programming. J Acoust Soc Japan 27(9):483–500

    Google Scholar 

  17. Sakoe H, Chiba S (1978) Dynamic programming algorithm optimization for spoken word recognition. IEEE Trans on Acoustics, Speech, and Signal Processing 26:43–49

    Article  MATH  Google Scholar 

  18. Shen J, Mei T, Gao X (2010) Automatic video archaeology: tracing your online videos. In: Workshop of social media, in conjunction of ACM MM, pp 59–64

  19. Silverman H, Morgan DP (1982) The application of dynamic programming to cconnected speech recognition. In: Proc. of international conference on acoustics, speech, and signal processing, pp 1649–1652

  20. Videosurf’s new release. available: http://www.videosurf.com/blog/videosurfs-new-release-1075/. Accessed Oct 2011

  21. Wang Z, Sheikh HR, Bovik AC (2002) No-reference pperceptual qquality assessment of jpeg ccompressed images. In: Proc. of IEEE international conference on image processing, vol 20, pp 477–480

  22. Wu X, Hauptmann AG, Ngo CW (2007) Practical elimination of near-duplicates from web video search. In: Proc. of the 15th international conference on multimedia, pp 218–227

  23. Yahoo! video. Available: http://video.yahoo.com/. Accessed Oct 2011

  24. Youtube. Available: http://www.youtube.com/. Accessed Oct 2011

  25. Youssef AM, Abdel-Galil TK, El-Saadany EF, Salama MMA (2004) Disturbance classification utilizing dynamic time warping classifier. IEEE Trans Power Deliv 19(1):272–278

    Article  Google Scholar 

  26. Yuan J, Duan L, Tian Q, Xu C (2004) Fast and robust short video clip search using an index structure, pp 61–68

  27. Zhang H-J, Kankanhalli A, Smoliar SW (1993) Automatic partitioning of full-motion video. Multimedia Syst 1(1):10–28

    Article  Google Scholar 

  28. Zhao W, Ngo CW, Tan H, Wu X (2007) Near-duplicate keyframe identification with interest point matching and pattern learning. IEEE Trans Multimedia 9(5):1037–1048

    Article  Google Scholar 

  29. Zhou X, Zhou X, Chen L (2009) An efficient near-duplicate video shot detection method using shot-based interest points. IEEE Trans Multimedia 11(5):879–891

    Article  Google Scholar 

Download references

Acknowledgements

We want to thank the helpful comments and suggestions from the anonymous reviewers. This research was supported partially by the National Natural Science Foundation of China under Grants 61125204, 61172146, 60832005 and 41031064, the Ph.D. Programs Foundation of Ministry of Education of China under Grant 20090203110002, and the Research Projector Funding of Microsoft Research Asia.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tao Mei.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Shen, J., Mei, T. & Gao, X. Video archaeology: understanding video manipulation history. Multimed Tools Appl 63, 461–483 (2013). https://doi.org/10.1007/s11042-011-0922-y

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-011-0922-y

Keywords

Navigation