Abstract
Visual tracking is one of hot researches in computer vision in recent years. C-COT [8] has obtained excellent results on many visual tracking benchmarks. However, it cannot exploit CNN features effectively because it gave the same weight for different CNN features. Furthermore, it updated model frame by frame, it possibly results in model drift. To address these problems, we propose an improved C-COT based visual tracking scheme to weighted fusion of diverse features. We present a weighted sum model that convolutional responses from different convolutional layers are weighted and summed to obtain the final response score. Secondly, we introduce a context based updating strategy for high confidence model update to avoid samples corruption and model drift. The experimental results on the challenging OTB dataset demonstrate that the proposed method is more competitive than state-of-the-art trackers.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Wu, Y., Lim, J., Yang, M.H.: Object tracking benchmark. TPAMI 37(9), 1834–1848 (2015)
Kristan, M., Pflugfelder, R., Leonardis, A., Matas, J., et al.: The visual object tracking VOT2014 challenge results. In: Agapito, L., Bronstein, M.M., Rother, C. (eds.) ECCV 2014. LNCS, vol. 8926, pp. 191–217. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-16181-5_14
Bolme, D.S., Beveridge, J.R., Draper, B.A., Lui, Y.M.: Visual object tracking using adaptive correlation filters. In: IEEE Conference on CVPR, pp. 2544–2550 (2010)
Danelljan, M., Shahbaz Khan, F., Felsberg, M., Van de Weijer, J.: Adaptive color attributes for real-time visual tracking. In: CVPR, pp. 1090–1097 (2014)
Danelljan, M., Hager, G., Khan, F., Felsberg, M.: Accurate scale estimation for robust visual tracking. In: BMVC Press (2014)
Danelljan, M., Hager, G., Shahbaz Khan, F., Felsberg, M.: Learning spatially regularized correlation filters for visual tracking. In: ICCV, pp. 4310–4318 (2015)
Wang, N., Shi, J., Yeung, D.-Y., Jia, J.: Understanding and diagnosing visual tracking systems. In: ICCV, pp. 3101–3109 (2015)
Danelljan, M., Robinson, A., Shahbaz Khan, F., Felsberg, M.: Beyond correlation filters: learning continuous convolution operators for visual tracking. In: Leibe, Bastian, Matas, Jiri, Sebe, Nicu, Welling, Max (eds.) ECCV 2016. LNCS, vol. 9909, pp. 472–488. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46454-1_29
Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.S.: Fully-convolutional siamese networks for object tracking. In: Hua, G., Jégou, H. (eds.) ECCV 2016. LNCS, vol. 9914, pp. 850–865. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-48881-3_56
Nam, H., Han, B.: Learning multi-domain convolutional neural networks for visual tracking. In: CVPR, pp. 4293–4302 (2015)
Wang, M., Liu, Y., Huang, Z.: Large margin object tracking with circulant feature maps. In: CVPR, pp. 4800–4808 (2017)
Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: ECO: efficient convolution operators for tracking. arXiv preprint arXiv:1611.09224 (2016)
Fan, H., Ling, H.: SANet: structure-aware network for visual tracking. In: Computer Vision and Pattern Recognition Workshops, pp. 2217–2224. IEEE (2017)
Danelljan, M., Hager, G., Shahbaz Khan, F., Felsberg, M.: Convolutional features for correlation filter based visual tracking. In: ICCV Workshops, pp. 58–66 (2015)
Zhang, T., Xu, C., Yang, M.H.: Multi-task correlation particle filter for robust object tracking. In: CVPR, pp. 4819–4827 (2017)
Choi, J., Chang, H.J., Yun, S., et al.: Attentional correlation filter network for adaptive visual tracking. In: CVPR, pp. 4828–4837 (2017)
Bertinetto, L., Valmadre, J., Golodetz, S., et al.: Staple: complementary learners for real-time tracking. In: Computer Vision and Pattern Recognition, pp. 1401–1409. IEEE (2016)
Qi, Y., Zhang, S., Qin, L., et al.: Hedged deep tracking. In: Computer Vision and Pattern Recognition, pp. 4303–4311. IEEE (2016)
Acknowledgement
We thank Beijing National Natural Science Foundation of China(61702022), Postdoctoral Research Foundation(2017-KZ-029), China Postdoctoral Science Foundation funded project(2018T110019), and Beijing university of technology “Ri Xin” cultivation project for partially supporting this work
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Wu, L., Wang, Q., Xu, D., Jian, M. (2018). An Improved C-COT Based Visual Tracking Scheme to Weighted Fusion of Diverse Features. In: Hong, R., Cheng, WH., Yamasaki, T., Wang, M., Ngo, CW. (eds) Advances in Multimedia Information Processing – PCM 2018. PCM 2018. Lecture Notes in Computer Science(), vol 11165. Springer, Cham. https://doi.org/10.1007/978-3-030-00767-6_63
Download citation
DOI: https://doi.org/10.1007/978-3-030-00767-6_63
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-00766-9
Online ISBN: 978-3-030-00767-6
eBook Packages: Computer ScienceComputer Science (R0)