Skip to main content

Learning Target-Specific Response Attention for Siamese Network Based Visual Tracking

  • Conference paper
  • First Online:
  • 1438 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 12002))

Abstract

Recently, the Siamese network based visual tracking methods have shown great potentials in balancing the tracking accuracy and computational efficiency. These methods use two-branch convolutional neural networks (CNNs) to generate a response map between the target exemplar and each of candidate patches in the search region. However, since these methods have not fully exploit the target-specific information contained in the CNN features during the computation of the response map, they are less effective to cope with target appearance variations and background clutters. In this paper, we propose a Target-Specific Response Attention (TSRA) module to enhance the discriminability of these methods. In TSRA, a channel-wise cross-correlation operation is used to produce a multi-channel response map, where different channels correspond to different semantic information. Then, TSRA uses an attention network to dynamically re-weight the multi-channel response map at every frame. Moreover, we introduce a shortcut connection strategy to generate a residual multi-channel response map for more discriminative tracking. Finally, we integrate the proposed TSRA into the classical Siamese based tracker (i.e., SiamFC) to propose a new tracker (called TSRA-Siam). Experimental results on three popular benchmark datasets show that the proposed TSRA-Siam outperforms the baseline tracker (i.e., SiamFC) by a large margin and obtains competitive performance compared with several state-of-the-art trackers.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Bertinetto, L., Valmadre, J., Golodetz, S., Miksik, O., Torr, P.H.S.: Staple: Complementary learners for real-time tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1401–1409 (2016)

    Google Scholar 

  2. Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.S.: Fully-convolutional siamese networks for object tracking. In: Hua, G., Jégou, H. (eds.) ECCV 2016. LNCS, vol. 9914, pp. 850–865. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-48881-3_56

    Chapter  Google Scholar 

  3. Cho, K., van Merrienboer, B., Gülçehre, Ç., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: Proceedings of Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1724–1734 (2014)

    Google Scholar 

  4. Choi, J., Chang, H.J., Fischer, T., Yun, S., Lee, K., Jeong, J., Demiris, Y., Choi, J.Y.: Context-aware deep feature compression for high-speed visual tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 479–488 (2018)

    Google Scholar 

  5. Choi, J., Chang, H.J., Yun, S., Fischer, T., Demiris, Y., Choi, J.Y.: Attentional correlation filter network for adaptive visual tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4828–4837 (2017)

    Google Scholar 

  6. Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1800–1807 (2017)

    Google Scholar 

  7. Cui, Z., Xiao, S., Feng, J., Yan, S.: Recurrently target-attending tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1449–1458 (2016)

    Google Scholar 

  8. Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: ECO: efficient convolution operators for tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6931–6939 (2017)

    Google Scholar 

  9. Danelljan, M., Häger, G., Khan, F.S., Felsberg, M.: Learning spatially regularized correlation filters for visual tracking. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 4310–4318 (2015)

    Google Scholar 

  10. Dong, X., Shen, J.: Triplet loss in siamese network for object tracking. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 472–488. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01261-8_28

    Chapter  Google Scholar 

  11. Fan, J., Wu, Y., Dai, S.: Discriminative spatial attention for robust tracking. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6311, pp. 480–493. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15549-9_35

    Chapter  Google Scholar 

  12. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778 (2016)

    Google Scholar 

  13. Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 37(3), 583–596 (2015)

    Article  Google Scholar 

  14. Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7132–7141 (2018)

    Google Scholar 

  15. Kalboussi, R., Abdellaoui, M., Douik, A.: Detecting and recognizing salient object in videos. In: Blanc-Talon, J., Helbert, D., Philips, W., Popescu, D., Scheunders, P. (eds.) ACIVS 2018. LNCS, vol. 11182, pp. 62–73. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01449-0_6

    Chapter  Google Scholar 

  16. Kristan, M., Leonardis, A., et al.: The visual object tracking VOT2017 challenge results. In: Proceedings of International Conference on Computer Vision Workshops (ICCV Workshops), pp. 1949–1972 (2017)

    Google Scholar 

  17. Li, B., Yan, J., Wu, W., Zhu, Z., Hu, X.: High performance visual tracking with siamese region proposal network. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8971–8980 (2018)

    Google Scholar 

  18. Lukezic, A., Vojir, T., Zajc, L.C., Matas, J., Kristan, M.: Discriminative correlation filter with channel and spatial reliability. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4847–4856 (2017)

    Google Scholar 

  19. Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vision 115(3), 211–252 (2015)

    Article  MathSciNet  Google Scholar 

  20. Tao, R., Gavves, E., Smeulders, A.W.M.: Siamese instance search for tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1420–1429 (2016)

    Google Scholar 

  21. Valmadre, J., Bertinetto, L., Henriques, J.F., Vedaldi, A., Torr, P.H.S.: End-to-end representation learning for correlation filter based tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5000–5008 (2017)

    Google Scholar 

  22. Wu, Y., Lim, J., Yang, M.: Online object tracking: a benchmark. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2411–2418 (2013)

    Google Scholar 

  23. Wu, Y., Lim, J., Yang, M.: Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1834–1848 (2015)

    Article  Google Scholar 

  24. Xu, K., Ba, J., et al.: Show, attend and tell: Neural image caption generation with visual attention, pp. 2048–2057 (2015). Computer Science

    Google Scholar 

  25. Zhang, Z., Peng, H.: Deeper and wider siamese networks for real-time visual tracking. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4591–4600 (2019)

    Google Scholar 

  26. Zhou, B., Khosla, A., Lapedriza, À., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2921–2929 (2016)

    Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No. U1605252 and 61872307) and the National Key R&D Program of China (Grant No. 2017YFB1302400).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hanzi Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhao, P., Chen, H., Liang, Y., Yan, Y., Wang, H. (2020). Learning Target-Specific Response Attention for Siamese Network Based Visual Tracking. In: Blanc-Talon, J., Delmas, P., Philips, W., Popescu, D., Scheunders, P. (eds) Advanced Concepts for Intelligent Vision Systems. ACIVS 2020. Lecture Notes in Computer Science(), vol 12002. Springer, Cham. https://doi.org/10.1007/978-3-030-40605-9_47

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-40605-9_47

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-40604-2

  • Online ISBN: 978-3-030-40605-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics