Skip to main content
Log in

3D video components generation using object tracking technique

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

2D-to-3D conversion that would be a solution of the lack of 3D contents has been a worthy and challenging research field. In this paper, we propose a computer interactive conversion method to capture components which is used to generate 3D sequences. First, we divide the key frame into foreground and background, and then label the objects by convenient computer interactive operation. Depth information of objects is labeled after segmentation. Second, we use object tracking technique which synthesizes the advantages of kernel-based mean shift tracker and contour tracker to accomplish object depth capture for non-key frame. Finally, all the 3D information is prepared to render 3D sequences. After all, we propose our future work direction: a 2D-to-3D system which can generate 3D sequence interactively.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

References

  1. Bai X, Wang J, Simons D, Sapiro G (2009) Video snapcut: robust video object cutout using localized classifiers. ACM Siggraph

  2. Boykov Y, Veksler O, Zabih R (2001) Fast approximate energy minimisation via graph cuts. IEEE Trans Pattern Anal Mach Intell 29:1222–1239

    Article  Google Scholar 

  3. Chen WY, Chang YL, Lin SF, Ding LF, Chen LG (2005) Efficient depth image based rendering with edge dependent depth filter and interpolation. In: IEEE International conference on multimedia and expo, ICME, pp 1314–1317

  4. Cheng CC, Li CT, Huang PS, Lin TK, Tsai YM, Chen LG (2009) A block-based 2D-to-3D conversion system with bilateral filter. In: 2009 Digest of technical papers international conf. consu. elec. ICCE, pp 10–14

  5. Cheng CC, Li CT, Tsai YM, Chen LG (2009) Hybrid depth cueing for 2D-to-3D conversion system. In: Proc. SPIE, vol 7237

  6. Fehn C (2004) Depth-image-based rendering(DIBR), compression, and transmission for a new approach on 3DTV. In: Proc. SPIE, vol 5291

  7. Goesele M, Snavely N, Curless B, Hoppe H, Seitz SM (2007) Multiview stereo for community photo collections. In: International conference on computer vision, ICCV, pp 1–8

  8. Guttmann M, Wolf L, Cohen-or D (2009) Semi-automatic stereo extraction from video footage. In: International conference on computer vision, ICCV, pp 136–142

  9. Harman P, Flack J, Fox S, Dowley M (2002) Rapid 2D to 3D Conversion. In: 13th Annual stereoscopic displays and applications conference 9th annual engineering reality of virtual reality conference, vol 4660, pp 78–86

  10. INRIA Labs (2004) http://groups.inf.ed.ac.uk/vision/CAVIAR/CAVIARDATA1/

  11. Kim J, Choe Y, Kim YG (2011) Robust MRF-based object tracking and graph-cut-based contour refinement for high quality 2D to 3D video conversion. In: 2011 IEEE Pacific rim conf. communications. Computers and Signal Processing, pp 358–363

  12. Kim D, Min D, Sohn K (2008) A stereoscopic video generation method using stereoscopic display characterization and motion analysis. IEEE Trans Broadcasting 54(2):188–197

    Article  Google Scholar 

  13. Kim M, Park S, Kim H, Artem I (2005) Automatic conversion of two-dimensional video into stereoscopic video. In: Proc. SPIE, vol 6016

  14. Kim M, Song M, Kim D, Choi K (1998) Stereoscopic conversion of monoscopic video by the transformation of vertical to horizontal disparity. In: Proc. SPIE, vol 3295

  15. Knorr S, Imre E, Ozkalayci B, Alatan AA, Sikora T (2006) A modular scheme for 2D/3D conversion of TV broadcast. In: Proc. 3DPVT

  16. Li Y, Sun J, Tang CK, Shum HY (2004) Lazy snapping. ACM Trans Graphics 23:303–308

    Article  Google Scholar 

  17. Li G, Wu H (2011) Robust object tracking using Kernel-based weighted fragments. In: Multimedia technology international conference, pp 3643–3646

  18. Lowe DG (1999) Object recognition from local scale-invariant features. In: International conference on computer vision, ICCV, vol 2, pp 1150–1157

  19. Matsumoto Y, Terasaki H, Sugimoto K, Arakawa T (1997) Conversion system of monocular image sequence to stereo using motion parallax. In: Stereoscopic displays and virtual reality systems, SPIE, vol 108

  20. Pollefeys M, Van Gool L, Vergauwen M, Verbiest F, Cornelis K, Tops J, Koch R (2004) Visual modeling with a hand-held camera. Int J Comput Vis 59(3):207–232

    Article  Google Scholar 

  21. Rotem E, Wolowelsky K, Pelz D (2005) Automatic video to stereoscopic video conversion. In: Proc. SPIE, vol 5664

  22. Saxena A, Sun M, Ng AY (2007) Learning 3-D scene structure from a single still image. In: International conference on computer vision, ICCV, pp 1–8

  23. Tam WJ, Zhang L (2006) 3D-TV content generation: 2D-3D conversion. In: IEEE International conference on multimedia and expo, pp 1869–1872

  24. Tsai YM, Chang YL, Chen LG (2006) Block-based vanishing line and vanishing point detection for 3D scene reconstruction. In: 2006 international symposium on intelligent signal processing and communications, vol 1 and 2, pp 541–544

  25. Xu F, Er G, Xie X, Dai Q (2008) 2D-to-3D conversion based on motion and color mergence. In: IEEE 3DTV Conf.

  26. Yilmaz A, Li X, Shah U (2004) Contour-based object tracking with occlusion handling in video acquired sing mobile cameras. IEEE Trans Pattern Anal Mach Intell 26

  27. Zhang L, Lawrence B, Wang D, Vincent A (2005) Comparison study on feature matching and block matching for automatic 2D-to-3D video conversion. In: IEEE. Conf. Visual Media Prod., pp 122–129

  28. Zhang L, Vzquez C, Knorr S (2011) 3DTV content creation: automatic 2D-to-3D video conversion. IEEE Trans Broadcast 57

Download references

Acknowledgements

This research is supported by the National Key Basic Research and Development Program of China (973)(No. 2013CB329505), the National Natural Science Foundation of China (61232011), NSFC—Guangdong Joint Fund (No. U0935004, U1201252), the National Key Technology R&D Program (No. 2011BAH27B01, 2011BHA16B08).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Shuxu Guo or Siming Meng.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Jiang, H., Guo, S., Meng, S. et al. 3D video components generation using object tracking technique. Multimed Tools Appl 71, 435–449 (2014). https://doi.org/10.1007/s11042-013-1451-7

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-013-1451-7

Keywords

Navigation