Skip to main content
Log in

Robust visual tracking based on spatial context pyramid

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

In recent years, discriminative correlation filter (DCF) has gained a lot of popularity in visual tracking, mainly due to its circular sampling from limited training data and computational efficiency in Fourier domain. However, such trackers do not make reasonable use of context information, resulting in limited performance. In this paper, we propose a novel DCF tracking framework based on spatial context pyramid (SCPT) to overcome this problem. Firstly, we take global spatial context into account to exploit the relationship between the target and its context for better tracking. Secondly, we design an effective spatial window to highlight the target while suppressing the background, and thus a robust filter model which has a high response for the target and low response for the background can be learned. Thirdly, we construct a context pyramid representation using multi-level spatial windows for adapting different challenging factors. To validate the compatibility of the proposed algorithm, we implement two versions with the representations from both conventional features and deep convolutional neural network (CNN) features. Extensive experimental results on the OTB-2013 benchmark demonstrate the effectiveness of the proposed tracker in terms of accuracy and robustness.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Babenko B, Yang MH, Belongie S (2011) Robust object tracking with online multiple instance learning. IEEE Trans Pattern Anal Mach Intell 33(8):1619–1632

    Article  Google Scholar 

  2. Bertinetto L, Valmadre J, Henriques JF et al (2016) Fully-convolutional siamese networks for object tracking. In: European conference on computer vision, pp 850–865

  3. Bolme DS, Beveridge JR, Draper BA et al (2010) Visual object tracking using adaptive correlation filters. In: IEEE conference on computer vision and pattern recognition, pp 2544–2550

  4. Chen X, He F, Yu H (2018) A matting method based on full feature coverage. Multimed Tools Appl 1–29. https://doi.org/10.1007/s11042-018-6690-1

  5. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: IEEE conference on computer vision and pattern recognition, vol 1, pp 886–893

  6. Danelljan M, Häger G, Khan F et al (2014) Accurate scale estimation for robust visual tracking. In: British machine vision conference

  7. Danelljan M, Shahbaz Khan F, Felsberg M et al (2014) Adaptive color attributes for real-time visual tracking. In: IEEE conference on computer vision and pattern recognition, pp 1090–1097

  8. Dinh TB, Vo N, Medioni G (2011) Context tracker: exploring supporters and distracters in unconstrained environments. In: IEEE conference on computer vision and pattern recognition, pp 1177–1184

  9. Galoogahi HK, Fagg A, Lucey S (2017) Learning background-aware correlation filters for visual tracking. In: IEEE international conference on computer vision, pp 1144–1152

  10. Hare S, Golodetz S, Saffari A et al (2016) Struck: structured output tracking with kernels. IEEE Trans Pattern Anal Mach Intell 38(10):2096–2109

    Article  Google Scholar 

  11. Henriques JF, Caseiro R, Martins P et al (2012) Exploiting the circulant structure of tracking-by-detection with kernels. In: European conference on computer vision, pp 702–715

  12. Henriques JF, Caseiro R, Martins P et al (2015) High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Mach Intell 37(3):583–596

    Article  Google Scholar 

  13. Hong S, You T, Kwak S et al (2015) Online tracking by learning discriminative saliency map with convolutional neural network. In: International conference on machine learning, pp 597–606

  14. Jia X, Lu H, Yang MH (2012) Visual tracking via adaptive structural local sparse appearance model. In: IEEE conference on computer vision and pattern recognition, pp 1822–1829

  15. Kalal Z, Mikolajczyk K, Matas J (2012) Tracking-learning-detection. IEEE Trans Pattern Anal Mach Intell 34(7):1409

    Article  Google Scholar 

  16. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, pp 1097–1105

  17. Li X, Hu W, Shen C et al (2013) A survey of appearance models in visual object tracking. ACM Trans Intell Syst Technol (TIST) 4(4):58

    Google Scholar 

  18. Li Y, Zhu J, Hoi SCH (2015) Reliable patch trackers: robust visual tracking by exploiting reliable patches. In: IEEE conference on computer vision and pattern recognition, pp 353–361

  19. Li K, He F, Yu H et al (2017) A parallel and robust object tracking approach synthesizing adaptive Bayesian learning and improved incremental subspace learning. Frontiers of Computer Science 1–20. https://doi.org/10.1007/s11704-018-6442-4

  20. Li K, He F, Yu H et al (2017) A correlative classifiers approach based on particle filter and sample set for tracking occluded target. Appl Math—J Chin Univ 32 (3):294–312

    Article  MathSciNet  Google Scholar 

  21. Li K, He F Z, Yu HP (2018) Robust Visual Tracking based on convolutional features with illumination and occlusion handing. J Comput Sci Technol 33(1):223–236

    Article  Google Scholar 

  22. Lienhart R, Maydt J (2002) An extended set of haar-like features for rapid object detection. In: IEEE international conference on image processing

  23. Liu S, Zhang T, Cao X et al (2016) Structural correlation filter for robust visual tracking. In: IEEE conference on computer vision and pattern recognition, pp 4312–4320

  24. Lukezic A, Vojir T, Zajc L C et al (2017) Discriminative correlation filter with channel and spatial reliability. In: IEEE conference on computer vision and pattern recognition, pp 4847–4856

  25. Ma C, Yang X, Zhang C et al (2015) Long-term correlation tracking. In: IEEE conference on computer vision and pattern recognition, pp 5388–5396

  26. Mueller M, Smith N, Ghanem B (2017) Context-aware correlation filter tracking. In: IEEE conference on computer vision and pattern recognition, pp 1387–1395

  27. Naresh Boddeti V, Kanade T, Vijaya Kumar BVK (2013) Correlation filters for object alignment. In: IEEE conference on computer vision and pattern recognition, pp 2291–2298

  28. Ni B, Kassim AA, Winkler S (2008) A hybrid framework for 3-D human motion tracking. IEEE Trans Circuits Syst Video Technol 18(8):1075–1084

    Article  Google Scholar 

  29. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556

  30. Smeulders AWM, Chu DM, Cucchiara R et al (2014) Visual tracking: an experimental survey. IEEE Trans Pattern Anal Mach Intell 36(7):1442–1468

    Article  Google Scholar 

  31. Vedaldi A, Lenc K (2015) Matconvnet: convolutional neural networks for matlab. In: Proceedings of the 23rd ACM international conference on multimedia, pp 689–692

  32. Walia GS, Kapoor R (2016) Robust object tracking based upon adaptive multi-cue integration for video surveillance. Multimed Tools Appl 75(23):15821–15847

    Article  Google Scholar 

  33. Wang N, Yeung DY (2013) Learning a deep compact image representation for visual tracking. In: Advances in neural information processing systems, pp 809–817

  34. Wang N, Li S, Gupta A et al (2015) Transferring rich feature hierarchies for robust visual tracking. arXiv preprint arXiv:1501.04587

  35. Wang L, Ouyang W, Wang X et al (2015) Visual tracking with fully convolutional networks. In: IEEE international conference on computer vision, pp 3119–3127

  36. Wang Z, Wang H, Tan J et al (2017) Robust object tracking via multi-scale patch based sparse coding histogram. Multimed Tools Appl 76(10):12181–12203

    Article  Google Scholar 

  37. Wang M, Liu Y, Huang Z (2017) Large margin object tracking with circulant feature maps. In: IEEE conference on computer vision and pattern recognition, pp 21–26

  38. Wen L, Cai Z, Lei Z et al (2012) Online spatio-temporal structural context learning for visual tracking. In: European conference on computer vision, pp 716–729

  39. Wu Y, Lim J, Yang MH (2013) Online object tracking: a benchmark. In: IEEE conference on computer vision and pattern recognition, pp 2411–2418

  40. Yu H, He F, Pan Y (2018) A novel region-based active contour model via local patch similarity measure for image segmentation. Multimed Tools Appl 77 (18):24097–24119

    Article  Google Scholar 

  41. Yu H, He F, Pan Y (2018) A novel segmentation model for medical images with intensity inhomogeneity based on adaptive perturbation. Multimed Tools Appl 1–20. https://doi.org/10.1007/s11042-018-6735-5

  42. Zhang K, Zhang L, Liu Q et al (2014) Fast visual tracking via dense spatio-temporal context learning. In: European conference on computer vision, pp 127–141

  43. Zhang H, Zhang X, Wang Y et al (2018) An experimental comparison of swarm optimization based abrupt motion tracking methods. IEEE Access 6:75383–75394

    Article  Google Scholar 

  44. Zhou Z, Zhou M, Shi X (2016) Target tracking based on foreground probability. Multimed Tools Appl 75(6):3145–3160

    Article  Google Scholar 

  45. Zhou Y, He F, Qiu Y (2017) Dynamic strategy based parallel ant colony optimization on GPUs for TSPs. Sci China Inf Sci 60(6):068102

    Article  Google Scholar 

  46. Zhou Y, He F, Hou N et al (2018) Parallel ant colony optimization on multi-core SIMD CPUs. Futur Gener Comput Syst 79:473–487

    Article  Google Scholar 

  47. Zhong W, Lu H, Yang MH (2014) Robust object tracking via sparse collaborative appearance model. IEEE Trans Image Process 23(5):2356–2368

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China (61773262, 61503173), China Aviation Science Foundation (20142057006).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shiqiang Hu.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tang, F., Zhang, X., Lu, X. et al. Robust visual tracking based on spatial context pyramid. Multimed Tools Appl 78, 21065–21084 (2019). https://doi.org/10.1007/s11042-019-7416-8

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-019-7416-8

Keywords

Navigation