Skip to main content

Multi-target Tracking with EmbedMask and LSTM Model Fusion

  • Conference paper
  • First Online:
  • 1004 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 12383))

Abstract

The problem of occlusion occurs during multi-target tracking may result in loss of characteristics of tracking target and thus lose the tracking targets. This paper proposes a multi-target vehicle tracking algorithm based on fusion of Embedding Coupling for One-stage Instance Segmentation (EmbedMask) and Long Short-Term Memory (LSTM) model. Firstly, the obtained real-time video data is input into EmbedMask target detection model by frame for target detection. The targets are separated from background, and traditional rectangular box detection is replaced by instance segmentation. Secondly, the maximum feature data of targets is generated by the resent convolution network, which is input into the LSTM model. The continuous data of targets is obtained by calculating and estimating the motion attitude of the tracking target. Finally, the motion and detection data of targets is input into new LSTM model layer, and the fusion calculation is used to reduce the tracking loss caused by overlapping, which can ensure the accuracy of target tracking. Experimental results on standard MOT data sets show that the proposed algorithm is robust and can be used to accurately track occluded overlapping targets.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Li, K., He, F., et al.: A parallel and robust object tracking approach synthesizing adaptive bayesian learning and improved incremental subspace learning. Front. Comput. Sci. 13, 1116–1135 (2019)

    Article  Google Scholar 

  2. Zhong, B., Yao, H., Chen, S., et al.: Visual tracking via weakly supervised learning from multiple imperfect oracles. Elsevier Pattern Recogn. 47(3), 1395–1410 (2014)

    Article  Google Scholar 

  3. Li, B., Yan, J., Wu, W., et al.: High performance visual tracking with siamese region proposal network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8971–8980 (2018)

    Google Scholar 

  4. Song, Y., Ma, C., et al.: Deep attentive tracking via reciprocative learning. In: Advances in Neural Information Processing Systems (NIPS), pp. 1931–1941 (2018)

    Google Scholar 

  5. Zhang, S., Qi, Y., Jiang, F., et al.: Point-to-set distance metric learning on deep representations for visual tracking. IEEE Trans. Intell. Transp. Syst. 19(1), 187–198 (2017)

    Article  Google Scholar 

  6. Wang, C., Huang, H., Zhang, T., Chen, J.: Limited memory eigenvector recursive principal component analysis in sensor-cloud based adaptive operational modal online identification. In: International Conference on Security, Privacy and Anonymity in Computation, Communication and Storage (SCS), pp. 119–129 (2019)

    Google Scholar 

  7. Bae, S.H., Yoon, K.J.: Confidence-based data association and discriminative deep appearance learning for robust online multi-object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 40(3), 595–610 (2018)

    Article  Google Scholar 

  8. Zhou, Q., Zhong, B., Zhang, Y., et al.: Deep Alignment network based multi-person tracking with occlusion and motion reasoning. IEEE Trans. Multimedia 21(5), 1183–1194 (2019)

    Article  Google Scholar 

  9. Chen, J., Sheng, H., Zhang, Y., Xiong, Z.: Enhancing detection model for multiple hypothesis tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR), pp. 18–27 (2017)

    Google Scholar 

  10. Choi, W.: Near-online multi-target tracking with aggregated local flow descriptor. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 3029–3037 (2015)

    Google Scholar 

  11. Chu, Q., Ouyang, W., Li, H., Wang, X., Liu, B., Yu, N.: Online multi-object tracking using cnn-based single object tracker with spatial-temporal attention mechanism. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 4836–4845 (2017)

    Google Scholar 

  12. Fang, K., Xiang, Y., Li, X., Savarese, S.: Recurrent autoregressive networks for online multi-object tracking. In: IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 466–475 (2018)

    Google Scholar 

  13. Fu, Z., Feng, P., Angelini, F., Chambers, J.A., Naqvi, S.M.: Particle PHD filter based multiple human tracking using online group-structured dictionary learning. IEEE Access 6, 14764–14778 (2018)

    Article  Google Scholar 

  14. Ying, H., Huang, Z., Liu, S., et al.: EmbedMask: embedding coupling for one-stage instance segmentation. arXiv preprint arXiv:1912.01954 (2019)

  15. Kim, C., Li, F., Rehg, J.M.: Multi-object tracking with neural gating using bilinear LSTM. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11212, pp. 208–224. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01237-3_13

    Chapter  Google Scholar 

  16. He, K., Gkioxari, G., Dollar, P., et al.: Mask R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2961–2969 (2017)

    Google Scholar 

  17. Tian, Z., Shen, C., Chen, H., et al.: FCOS: fully convolutional one-stage object detection. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 9627–9636 (2019)

    Google Scholar 

  18. Bolya, D., Zhou, C., Xiao, F., et al.: YOLACT: real-time instance segmentation. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 9157–9166 (2019)

    Google Scholar 

  19. Ma, Y., Zhu, X., Zhang, S., Yang, R., Wang, W., Manocha, D.: TrafficPredict: trajectory prediction for Heterogeneous traffic-agents. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pp. 6120–6127 (2019)

    Google Scholar 

  20. Milan, A., Leal-Taixé, L., Reid, I., Roth, S., Schindler, K.: MOT16: a benchmark for multi-object tracking (2016)

    Google Scholar 

  21. Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3354–3361 (2012)

    Google Scholar 

Download references

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China (Grants No. 61801323), by the Science and Technology Projects Fund of Suzhou (Grant No. SYG201708, Grant No. SS2019029), by the Construction System Science and Technology Fund of Jiangsu Province (Grant No. 2017ZD066).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chongben Tao .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tao, C., Lu, K., Cao, F. (2021). Multi-target Tracking with EmbedMask and LSTM Model Fusion. In: Wang, G., Chen, B., Li, W., Di Pietro, R., Yan, X., Han, H. (eds) Security, Privacy, and Anonymity in Computation, Communication, and Storage. SpaCCS 2020. Lecture Notes in Computer Science(), vol 12383. Springer, Cham. https://doi.org/10.1007/978-3-030-68884-4_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-68884-4_26

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-68883-7

  • Online ISBN: 978-3-030-68884-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics