skip to main content
10.1145/2168556.2168574acmconferencesArticle/Chapter ViewAbstractPublication PagesetraConference Proceedingsconference-collections
research-article

Multi-mode saliency dynamics model for analyzing gaze and attention

Published:28 March 2012Publication History

ABSTRACT

We present a method to analyze a relationship between eye movements and saliency dynamics in videos for estimating attentive states of users while they watch the videos. The multi-mode saliency-dynamics model (MMSDM) is introduced to segment spatio-temporal patterns of the saliency dynamics into multiple sequences of primitive modes underlying the saliency patterns. The MMSDM enables us to describe the relationship by the local saliency dynamics around gaze points, which is modeled by a set of distances between gaze points and salient regions characterized by the extracted modes. Experimental results show the effectiveness of the proposed model to classify the attentive states of users by learning the statistical difference of the local saliency dynamics on gaze-paths at each level of attentiveness.

References

  1. Brandherm, B., Prendinger, H., and Ishizuka, M. 2007. Interest estimation based on dynamic bayesian networks for visual attentive presentation agents. In Proc. of ICMI, 346--349. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Bregler, C. 1997. Learning and recognizing human dynamics in video sequences. In Proc. of CVPR, 568--574. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Calder, A., Lawrence, A., Keane, J., Scott, S., Owen, A., Christoffels, I., and Young, A. 2002. Reading the mind from eye gaze. Neuropsychologia 40, 8, 1129--1138.Google ScholarGoogle ScholarCross RefCross Ref
  4. Dorr, M., Martinetz, T., Gegenfurtner, K., and Barth, E. 2010. Variability of eye movements when viewing dynamic natural scenes. Journal of vision 10, 10, 1--17.Google ScholarGoogle ScholarCross RefCross Ref
  5. Doshi, A., and Trivedi, M. M. 2010. Attention estimation by simultaneous observation of viewer and view. In Proc. of CVPR Workshop, 21--27.Google ScholarGoogle Scholar
  6. Fletcher, L., and Zelinsky, A. 2009. Driver inattention detection based on eye gaze-road event correlation. International Journal of Robotics Research 28, 6, 774--801. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Goldstein, R. B., Woods, R. L., and Peli, E. 2007. Where people look when watching movies: do all viewers look at the same place? Computers in Biology and Medicine 37, 7, 957--64. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Hirayama, T., Dodane, J.-B., Kawashima, H., and Matsuyama, T. 2010. Estimates of user interest using timing structures between proactive content-display updates and eye movements. IEICE Trans. on Information and Systems E-93D, 6, 1470--1478.Google ScholarGoogle Scholar
  9. Itti, L., Koch, C., and Niebur, E. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. on PAMI 20, 11, 1254--1259. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Ji, Q., Lan, P., and Looney, C. 2006. A probabilistic framework for modeling and real-time monitoring human fatigue. IEEE Trans. on Systems, Man and Cybernetics, Part A: Systems and Humans 36, 5, 862--875. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Just, M., and Carpenter, P. 1976. Eye fixations and cognitive processes. Cognitive Psychology 480, 441--480.Google ScholarGoogle ScholarCross RefCross Ref
  12. Kahneman, D. 1973. Attention and effort. Prentice Hall.Google ScholarGoogle Scholar
  13. Kawashima, H., and Matsuyama, T. 2005. Multiphase learning for an interval-based hybrid dynamical system. IEICE Trans. on Fundamentals E88-A, 11, 3022--3035. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Li, Y., Wang, T., and Shum, H.-Y. 2002. Motion texture: a two-level statistical model for character motion synthesis. ACM Trans. on Graphics 21, 3, 465--472. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Munn, S. M., Stefano, L., and Pelz, J. B. 2008. Fixation-identification in dynamic scenes. In Proc. of APGV, 33--42. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Nakano, Y. I., and Ishii, R. 2010. Estimating user's engagement from eye-gaze behaviors in human-agent conversations. In Proc. of IUI, 139--148. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. North, B., Blake, A., Isard, M., and Rittscher, J. 2000. Learning and classification of complex dynamics. IEEE Trans. on PAMI 22, 9, 1016--1034. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Pavlovic, V., Rehg, J. M., and Maccormick, J. 2000. Learning switching linear models of human motion. In Proc. of NIPS, 981--987.Google ScholarGoogle Scholar
  19. Qvarfordt, P., and Zhai, S. 2005. Conversing with the user based on eye-gaze patterns. In Proc. of CHI, 221--230. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Schneider, W., and Shiffrin, R. M. 1977. Controlled and automatic human information processing: I. detection, search, and attention. Psychological Review 84, 1--66.Google ScholarGoogle ScholarCross RefCross Ref
  21. Walther, D., and Koch, C. 2006. Modeling attention to salient proto-objects. Neural Networks: the Official Journal of the International Neural Network Society 19, 9, 1395--1407. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Wildes, R. P. 1998. A measure of motion salience for surveillance applications. In Proc. of ICIP, 183--187.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Multi-mode saliency dynamics model for analyzing gaze and attention

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ETRA '12: Proceedings of the Symposium on Eye Tracking Research and Applications
      March 2012
      420 pages
      ISBN:9781450312219
      DOI:10.1145/2168556

      Copyright © 2012 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 28 March 2012

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate69of137submissions,50%

      Upcoming Conference

      ETRA '24
      The 2024 Symposium on Eye Tracking Research and Applications
      June 4 - 7, 2024
      Glasgow , United Kingdom

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader