Abstract
This paper concerns on overcoming the challenges caused by drastic appearance change in visual tracking, especially the long-term appearance variation due to occlusion or large object deformation. We aim to build a long-term appearance model for robust tracking against large appearance change in two new respects: using historical and distinguishing cues to model target representation and extracting effective spatial objectness features from each frame to distinguish outliers. For the first purpose, an adaptive superpixel-based appearance model is formulated. Different from previous superpixel-based trackers, a complementary feature set is defined for the update model to preserve the features of those temporally disappeared object parts especially under occlusion and large deformation. For the second purpose, three new spatial objectness cues specially designed for tracking are defined, including surrounding comparison, edge density change and weighted superpixel straddling. With these spatial objectness cues, our method facilitates target object localization and ensures the target has similar edge distribution between adjacent frames. These cues greatly improve the ability of our method to distinguish the target from its surrounding background. The adaptive appearance model retains valuable features of historical results, and the spatial objectness cues are extracted from the current frame, and thus they are finally combined to complement with each other to solve large appearance changes. The extensive evaluations on the CVPR 2013 online object tracking benchmark and VOT 2014 datasets demonstrate the effectiveness of our method as compared with related trackers.
Similar content being viewed by others
References
Smeulders AWM, Chu DM, Cucchiara R, Calderara S, Dehghan A, Shah M (2014) Visual tracking: an experiment survey. IEEE Trans Pattern Anal Mach Intell 36(7):1442–1468
Hong Z, Chen Z, Wang C, Mei X, Prokhorov D, Tao D (2015) MUlti-store tracker (MUSTer): a cognitive psychology inspired approach to object tracking. In: IEEE international conference on computer vision and pattern recognition, pp 749–758
Duffner S, Garcia C (2013) PixelTrack: a fast adaptive algorithm for tracking non-rigid objects. In: IEEE international conference on computer vision, pp 2480–2487
Bibi A, Mueller M, Ghanem B (2016) Target response adaptation for correlation filter tracking. In: European conference on computer vision (2016)
Ma C, Yang X, Zhang C, Yang MY (2015) Long-term correlation tracking. In: IEEE international conference on computer vision and pattern recognition, pp 5388–5396
Zhang S, Zhou H, Jiang F, Li X (2015) Robust visual tracking using structurally random projection and weighted least squares. IEEE Trans Circuits Syst Video Technol 25(11):1749–1760
Wang D, Lu H, Yang MH (2012) Online object tracking with sparse prototypes. IEEE Trans Image Process 22(1):314–325
Possegger H, Mauthner T, Bischof H (2015) In defense of color-based model-free tracking. In: IEEE international conference on computer vision and pattern recognition, pp 2113–2120
Danelljan M, Khan FS, Felsberg M, Weijer JV (2014) Adaptive color attributes for real-time visual tracking. In: IEEE international conference on computer vision and pattern recognition, pp 1090–1097
Wang S, Lu H, Yang F, Yang MH (2011) Superpixel tracking. In: IEEE international conference on computer vision, pp 1323–1330
Wen LY, Cai ZW, Lei Z, Yi D, Li SZ (2014) Robust online learned spatio-temporal context model for visual tracking. IEEE Trans Image Process 23(2):785–796
Kwon J, Roh J, Lee KM, Gool LV (2014) Robust visual tracking with double bounding box. In: European conference on computer vision, pp 377–392
Zhang K, Zhang L, Yang MH (2014) Fast compressive tracking. IEEE Trans Pattern Anal Mach Intell 36(10):2002–2015
Danelljan M, Hager G, Khan FS, Felsberg M (2016) Discriminative scale space tracking. IEEE Trans Pattern Anal Mach Intell 39:1561–1575
Li X, Dick A, Shen C, van den Hengel A, Wang H (2013) Incremental learning of 3D-DCT compact representations for robust visual tracking. IEEE Trans Pattern Anal Mach Intell 35(4):863–881
Bai QX, Wu Z, Sclaroff S, Betke M, Monnier C (2013) Randomized ensemble tracking. In: IEEE international conference on computer vision, pp 2040–2047
Santner J, Leistner C, Saffari A, Pock T, Bischof H (2010) PROST: Parallel robust online simple tracking. In: IEEE international conference on computer vision and pattern recognition, pp 723–730
Zhong W, Lu H, Yang MH (2012) Robust object tracking via sparsity based collaborative model. In: IEEE international conference on computer vision and pattern recognition, pp 1838–1845
Wang D, Lu HC (2014) Visual tracking via probability continuous outlier model. In: IEEE international conference on computer vision and pattern recognition, pp 3478–3485
Atkinson RC, Shiffrin RM (1968) Human memory: a proposed system and its control processes. Psychol Learn Motiv 2:89–195
Wang N, Shi J, Yeung DY, Jia J (2015) Understanding and diagnosing visual tracking systems. In: IEEE international conference on computer vision and pattern recognition, pp 3101–3109
Kwon J, Lee KM (2009) Tracking of a non-rigid object via patch-based dynamic appearance modeling and adaptive Basin Hopping Monte Carlo sampling. In: IEEE international conference on computer vision and pattern recognition, pp 1208–1215
Li C, Cheng H, Hu S, Liu X, Tang J, Lin L (2016) Learning collaborative sparse representation for grayscale-thermal tracking. IEEE Trans Image Process 25(12):5743–5756
Lan XY, Ma AJ, Yuen PC (2014) Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation. In: IEEE international conference on computer vision and pattern recognition, pp 1194–1201
Zhang T, Liu S, Xu C, Yan S, Ghanem B, Ahuja N, Yang MH (2015) Structural sparse tracking. In: IEEE international conference on computer vision and pattern recognition, pp 150–158
Chen DP, Yuan ZJ, Wu Y, Zhang G, Zheng NJ (2013) Constructing adaptive complex cells for robust visual tracking. In: IEEE international conference on computer vision, pp 1113–1120
Dai M, Cheng S, He X, Wang D (2018) Object tracking in the presence of shaking motions. Neural Comput Appl. https://doi.org/10.1007/s00521-018-3387-3
Li C, Lin L, Zuo W, Tang J, Yang MH (2018) Visual tracking via dynamic graph learning. In: IEEE transactions on pattern analysis and machine intelligence, pp 1–15
Zhong W, Lu HC, Yang MH (2012) Robust object tracking via sparsity-based collaborative model. In: IEEE international conference on computer vision and pattern recognition, pp 1838–1845
Sun S, An Z, Jiang X, Zhang B, Zhang J (2018) Robust object tracking with the inverse relocation strategy. Neural Comput Appl. https://doi.org/10.1007/s00521-018-3667-y
Choi J, Chang HJ, Fischer T, Yun S, Lee K, Jeong J, Demiris Y, Choi JY (2018) Context-aware deep feature compression for high-speed visual tracking. In: IEEE conference on computer vision and pattern recognition, pp 479–488
Jia X, Lu H, Yang MH (2012) Visual tracking via adaptive structural local sparse appearance model. In: IEEE international conference on computer vision and pattern recognition, pp 1822–1829
Hare S, Saffari A, Torr PHS (2011) Struck: structured output tracking with kernels. In: IEEE international conference on computer vision, pp 263–270
Li Y, Zhu J, Hoi SCH (2015) Reliable patch trackers: robust visual tracking by exploiting reliable patches. In: IEEE international conference on computer vision and pattern recognition, pp 353–361
Wang D, Lu H, Xiao Z, Yang MH (2015) Inverse sparse tracker with a locally weighted distance metric. IEEE Trans Image Process 24(9):2646–2657
Wang NY, Wang JD, Yeung DY (2013) Online robust non-negative dictionary learning for visual tracking. In: IEEE international conference on computer vision, pp 657–664
Henriques JF, Caseiro R, Martins P, Batista J (2015) High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Mach Intell 37(3):583–596
Godec M, Roth PM, Bischof H (2011) Hough-based tracking of non-rigid objects. In: IEEE international conference on computer vision, pp 81–88
Hu W, Zhou X, Hu M, Maybank S (2009) Occlusion reasoning for tracking multiple walking people. IEEE Trans Circuits Syst Video Technol 19(1):114–121
Hu M, Liu Z, Zhang J, Zhang G (2017) Robust object tracking via multi-cue fusion. Signal Process 139:86–95
Grabner H, Leistner C, Bischof H (2008) Semi-supervised on-line boosting for robust tracking. In: European conference on computer vision, pp 234–247
Zhang H, Cao X, Ho JKL, Chow TWS (2017) Object-level video advertising: an optimization framework. IEEE Trans Ind Inform 13:520–531
Kadir T, Zisserman A, Brady M (2004) An affine invariant salient region detector. In: European conference on computer vision, pp 228–241
Marchesotti L, Cifarelli C, Csurka G (2009) A framework for visual saliency detection with applications to image thumb nailing. In: IEEE international conference on computer vision, pp 2232–2239
Hou X, Zhang L (2007) Saliency detection: a spectral residual approach. In: IEEE international conference on computer vision and pattern recognition, pp 1–8
Alexe B, Deselaers T, Ferrari V (2012) Measuring the objectness of image windows. IEEE Trans Pattern Anal Mach Intell 34(11):2189–2202
Wu Y, Lim J, Yang MH (2013) Online object tracking: a benchmark. In: IEEE international conference on computer vision and pattern recognition, pp 2411–2418
Xing JL, Gao J, Li B, Hu WM, Yan SC (2013) Robust object tracking with online multi-lifespan dictionary learning. In: IEEE international conference on computer vision, pp 665–672
Radhakrishna A, Shaji A, Lucchi K, Fua P, Susstrunk S (2010) Slic-superpixels, No. EPFL-REPORT-149300
Kristan M, Pflugfelder R, et al (2014) The visual object tracking VOT2014 challenge results. In: European conference on computer vision (Workshop, 2014)
Henriques F, Caseiro R, Martins P, Batista J (2012) Exploiting the circulant structure of tracking-by-detection with kernels. In: European conference on computer vision (2012)
Acknowledgements
Many thanks go to the anonymous reviewers for their careful work and thoughtful suggestions that have helped us to achieve great and substantial improvement on this paper. This work was partially supported by the National Natural Science Fund of China (61772209, 61772257, 61672279), the Science and Technology Planning Project of Guangdong Province (2016A050502050). Wei-shi Zheng is the corresponding author of this paper.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
All authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
About this article
Cite this article
Liang, Y., Wang, Mh., Guo, Yw. et al. On large appearance change in visual tracking. Neural Comput & Applic 32, 6089–6109 (2020). https://doi.org/10.1007/s00521-019-04094-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-019-04094-z