Skip to main content
Log in

The invariant features-based target tracking across multiple cameras

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Target tracking across lenses is a popular research topic for video surveillance recently. This paper presents a method of target tracking across lenses with overlap regions. First, the target detection and tracking are completed with a single camera. Second, in order to obtain the location-invariant feature of the same target in the images with various cameras, the camera calibration is completed based on a three-dimension (3D) model. After that, for all images via multiple cameras, the coordinates of the 3D model are unified. Finally, referring to the assumption of spatial and temporal consistency of the target location across multiple cameras, the association among detected objects for the same target with different cameras is established. And a feature pool is built which contains perspective and scale features. Thus the same target is continuously tracked across multiple lenses. At last, the performance of the proposed approach is compared with KSP and PABC and demonstrated with indoor and outdoor experiments.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

References

  1. Berclaz J, Fleuret F, Turetken E, Fua P (2011) Multiple object tracking using Kshortest paths optimization. IEEE Trans Pattern Anal Mach Intell 33(9):1806–1819

    Article  Google Scholar 

  2. Black J, Ellis T Multiple camera image tracking, in: Proc. Performance Evaluation of Tracking and Surveillance Conference (PETS 2001), with CVPR 2001, December 2001.

  3. Bouchrika I, Carter JN, Nixon MS ( 2014) Towards automated visual surveillance using gait for identity recognition and tracking across multiple non-intersecting cameras. Multimed Tools Appl, pp. 1–21, doi: 10.1007/s11042-014-2364-9

  4. Calderara S, Cucchiara R, Prati A (2008) Bayesian-competitive consistent labeling for people surveillance. IEEE Trans PAMI 30(2):354–360

    Article  Google Scholar 

  5. Calderara S, Prati A, Vezzani, R (2005) Cucchiara, Consistent labeling for multi-camera object tracking, in: 13th International Conference on Image Analysis and Processing, September

  6. Chen X, Huang K, Tan T (2013) Object tracking across non-overlapping views by learning inter-camera transfer models. Pattern Recogn 47(3):1126–1137

    Article  Google Scholar 

  7. Chen K, Lai C, Hung Y, Chen C (2008) An adaptive learning method for target tracking across multiple cameras, in: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8

  8. Chen CH, Yao Y, Page D, Abidi B, Koschan A, Abidi M (2010) Camera handoff with adaptive resource management for multi-camera multi-object tracking. Image Vis Comput 28:851–864

    Article  Google Scholar 

  9. Comaniciu D, Ramesh V, Meer P (2000) Real-time tracking of non-rigid objects using mean Shift, BEST PAPER AWARD, IEEE Conf. Computer Vision and Pattern Recognition (CVPR’00), Hilton Head Island, South Carolina, Vol. 2, 142–149

  10. Finlayson GD, Trezzi E (2004) Shades of gray and colour constancy, in: Proceedings of the 12th Color Imaging Conference: Color Science and Engineering Systems, Technologies, and Applications, pp. 37–41

  11. Fluuret F, Berclaz J, Lengagne R, Fua P (2008) Multicamera people tracking with a probabilistic occupancy map. IEEE Trans PAMI 30(2):267–273

    Article  Google Scholar 

  12. Forsyth D (1990) A novel algorithm for color constancy. Int J Comput Vis 5(1):5–36

    Article  Google Scholar 

  13. Gijsenij A, Gevers T, van de Weijer J (2010) Generalized gamut mapping using image derivative structures for color constancy. Int J Comput Vis 86(2):127

    Article  Google Scholar 

  14. Gilbert A, Bowden R (2006) Tracking objects across cameras by incrementally learning inter-camera colour calibration and patterns of activity, in: European Conference on Computer Vision, pp. 125–136.

  15. Hu W, Hu M, Zhou X, Tan T, Lou J, Maybank S (2006) Principal axis-based correspondence between multiple cameras for people tracking. IEEE Trans PAMI 28(4):663–671

    Article  Google Scholar 

  16. Javed O, Shafique K, Rasheed Z, Shah M (2008) Modeling inter-camera space-time and appearance relationships for tracking across non-overlapping views. Comput Vis Image Underst 109:146–162

    Article  Google Scholar 

  17. Javed O, Shafique K, Shah M (2005) Appearance modeling for tracking in multiple non-overlapping cameras, in: Proceedings of the IEEE Computer Society Conference on Computer Vision and. Pattern Recogn 2:26–33

    Google Scholar 

  18. Jeong K, Jaynes C (2008) Object matching in disjoint cameras using a color transfer approach. Mach Vis Appl 19:443–455

    Article  MATH  Google Scholar 

  19. Kang J, Cohen I, Medion Gi (2003) Continuous tracking within and across camera streams, in: IEEE International Conference on Computer Vision and Pattern Recognition, June 2003

  20. Kelly P, Katkere A Kuramura D, Moezzi S, Chatterjee S Jain R (1995) An architecture for multiple perspective interactive video, in: Proceedings of ACM Multimedia 95, May

  21. Khan S, Shah M (2003) Consistent labeling of tracked objects in multiple cameras with overlapping fields of view. IEEE Trans PAMI 25(10):1355–1361

    Article  Google Scholar 

  22. Lee L, Romano R, Stein G (2000) Monitoring activities from multiple video streams: establishing a common coordinate frame. IEEE Trans PAMI 22(8):758–767

    Article  Google Scholar 

  23. Lian G, Lai J, Zheng WS (2011) Spatial-temporal consistent labelling of tracked pedestrians across non-overlapping camera views. Pattern Recogn 44(5):1121–1136

    Article  MATH  Google Scholar 

  24. Nam Y, Rho S, Park JH (2013) Inference topology of distributed camera networks with multiple cameras. Multimed Tools Appl 67(1):289–309

    Article  Google Scholar 

  25. Prosser B Gong S Xiang T (2008) Multi-camera matching using bi-directional cumulative brightness transfer functions, in: BMVC

  26. Stauffer C, Grimson WEL (1999) Adaptive background mixture models for real-time tracking. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Fort Collins, USA:IEEE, 246–252

Download references

Acknowledgments

This work has been sponsored, in part, by Beijing Nova Program (Z131101000413083), Beijing Talents Fund (2014000021223ZK41) and Aeronautic Science Foundation of China (2015ZC51032). All supports are gratefully acknowledged.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jin Xiao.

Ethics declarations

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xiao, J., Liu, Z., Yang, H. et al. The invariant features-based target tracking across multiple cameras. Multimed Tools Appl 76, 12165–12179 (2017). https://doi.org/10.1007/s11042-015-3067-6

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-015-3067-6

Keywords

Navigation