skip to main content
10.1145/1878083.1878089acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Virtual gazing in video surveillance

Published: 29 October 2010 Publication History

Abstract

Although a computer can track thousands of moving objects simultaneously, it often fails to understand the priority and the meaning of the dynamics. Human vision, on the other hand, can easily track multiple objects with saccadic motion. The single thread eye movement allows people to shift attention from one object to another, enabling visual intelligence from complex scenes. In this paper, we present a motion-context attention shift (MCAS) model to simulate attention shifts among multiple moving objects in surveillance videos. The MCAS model includes two modules: The robust motion detector module and the motion-saliency module. Experimental results show that the MCAS model successfully simulates the attention shift in tracking multiple objects in surveillance videos.

References

[1]
M. A. Just and P. A. Carpenter. A theory of reading: From eye fixations to comprehension. Psychological Review, vol. 87(4), pp. 329--354, 1980.
[2]
D. Beymer and D. M. Russell. Web gaze analyzer: a system for capturing and analyzing web reading behavior using eye gaze. In Proceedings of CHI'05, 2005, pp. 1913--1916.
[3]
R. J. K. Jacob. The use of eye movements in human computer interaction techniques: Toward non-command interfaces. ACM Transactions on Information Systems, vol. 9 (3), pp. 152--169, 1991.
[4]
H. Takagi. Development of an eye-movement enhanced translation support system. In Proc. Asian-Pacific Computer Human Interaction Conference (APCHI), 1998, pp. 114--119.
[5]
J. L. Sibert, M. Gokturk, and R. A. Lavine. The reading assistant: Eye gaze triggered auditory prompting for reading remediation. In Proceedings of CHI'07, 2000, pp. 101--107.
[6]
S. T. Iqbal and B. P. Bailey. Understanding and developing models for detecting and differentiating breakpoints during interactive tasks. In Proceedings of CHI'07, 2007.
[7]
Yan Liu, Pei-Yun Hsueh, Lai, J., Sangin, M., Nussli, M.-A., Dillenbourg, P. Who is the expert? Analyzing gaze data to predict expertise level in collaborative applications. Proc. of IEEE International Conference on Multimedia and Expo, 2009.
[8]
http://www.eyetechds.com/.
[9]
Wang Jian. Integration model of eye-gaze, voice and manual response in multimodal user interface. Journal of Computer Science and Technology, vol.11(5), pp. 512--518, 1996
[10]
Helena Grillon, Daniel Thalmann. Simulating gaze attention behaviors for crowds. Journal of Visualization and Computer Animation 20(2--3): 111--119 (2009).
[11]
Catherine Pelachaud, Massimo Bilvi: Modelling Gaze Behaviour for Conversational Agents. IVA 2003: 93--100.
[12]
L. Li, W. Huang, I. Y. H. Gu, Q. Tian, Foreground object detection from videos containing complex background, ACM Multimedia, 2003.
[13]
D. Comaniciu, V. Ramesh, P. Meer, Real-time tracking of non-rigid objects using mean shift, in: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR'00), South Carolina, 2000, pp. 142--149.
[14]
L. Itti, P. Baldi, A Principled Approach to Detecting Surprising Events in Video, In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 631--637, Jun 2005.
[15]
L. Itti, P. Baldi, Bayesian Surprise Attracts Human Attention, In Advances in Neural Information Processing Systems, Vol. 19 (NIPS*2005), pp. 1--8, Cambridge, MA: MIT Press, 2006.
[16]
W. Einhaeuser, T. N. Mundhenk, P. Baldi, C. Koch, L. Itti, A bottom-up model of spatial attention predicts human error patterns in rapid scene recognition, Journal of Vision, Vol. 7, No. 10, pp. 1--13, Jul 2007.
[17]
A. L. Yarbus, Eye Movements and Vision. New York: Plenum Press, 1967.
[18]
A. T. Duchowski, Eye Tracking Methodologies: Theory and Practice. Springer, 2002.
[19]
http://www.youtube.com/watch?v=47LCLoidJh4.
[20]
M. Yang, J. Yuan and Y. Wu. Spatial selection for attentional visual tracking, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR) pp. 1--8, 2007.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SMVC '10: Proceedings of the 2010 ACM workshop on Surreal media and virtual cloning
October 2010
76 pages
ISBN:9781450301756
DOI:10.1145/1878083
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 29 October 2010

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. motion detector
  2. motion-context attention shift
  3. motion-saliency module
  4. simulation
  5. virtual gazing

Qualifiers

  • Research-article

Conference

MM '10
Sponsor:
MM '10: ACM Multimedia Conference
October 29, 2010
Firenze, Italy

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 120
    Total Downloads
  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)0
Reflects downloads up to 14 Feb 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media