skip to main content
10.1145/2857491.2857515acmconferencesArticle/Chapter ViewAbstractPublication PagesetraConference Proceedingsconference-collections
research-article

Deep eye fixation map learning for calibration-free eye gaze tracking

Published:14 March 2016Publication History

ABSTRACT

The existing eye trackers typically require an explicit personal calibration procedure to estimate subject-dependent eye parameters. Despite efforts in simplifying the calibration process, such a calibration process remains unnatural and bothersome, in particular for users of personal and mobile devices. To alleviate this problem, we introduce a technique that can eliminate explicit personal calibration. Based on combining a new calibration procedure with the eye fixation prediction, the proposed method performs implicit personal calibration without active participation or even knowledge of the user. Specifically, different from traditional deterministic calibration procedure that minimizes the differences between the predicted eye gazes and the actual eye gazes, we introduce a stochastic calibration procedure that minimizes the differences between the probability distribution of the predicted eye gaze and the distribution of the actual eye gaze. Furthermore, instead of using saliency map to approximate eye fixation distribution, we propose to use a regression based deep convolutional neural network (RCNN) that specifically learns image features to predict eye fixation. By combining the distribution based calibration with the deep fixation prediction procedure, personal eye parameters can be estimated without explicit user collaboration. We apply the proposed method to both 2D regression-based and 3D model-based eye gaze tracking methods. Experimental results show that the proposed method outperforms other implicit calibration methods and achieve comparable results to those that use traditional explicit calibration methods.

References

  1. Alnajar, F., Gevers, T., Valenti, R., and Ghebreab, S. 2013. Calibration-free gaze estimation using human gaze patterns.Google ScholarGoogle Scholar
  2. Beymer, D., and Flickner, M. 2003. Eye gaze tracking using an active stereo head. IEEE Conference in Computer Vision and Pattern Recognition.Google ScholarGoogle Scholar
  3. Cerf, M., Harel, J., Einhäuser, W., and Koch, C. 2008. Predicting human gaze using low-level saliency combined with face detection. In Advances in neural information processing systems.Google ScholarGoogle Scholar
  4. Cerf, M., Frady, E. P., and Koch, C. 2009. Faces and text attract gaze independent of the task: Experimental data and computer model. Journal of vision.Google ScholarGoogle ScholarCross RefCross Ref
  5. Chen, J., and Ji, Q. 2011. Probabilistic gaze estimation without active personal calibration. IEEE Conference on Computer Vision and Pattern Recognition. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Chen, J., and Ji, Q. 2014. A probabilistic approach to online eye gaze tracking without personal calibration. IEEE Transactions on Image Processing.Google ScholarGoogle Scholar
  7. Chen, J., Tong, Y., Gary, W., and Ji, Q. 2008. A robust 3d eye gaze tracking system using noise reduction. Proceedings of the 2008 symposium on Eye tracking research and applications. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Einhäuser, W., Spain, M., and Perona, P. 2008. Objects predict fixations better than early saliency. Journal of Vision.Google ScholarGoogle ScholarCross RefCross Ref
  9. Fukushima, K. 1980. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biological cybernetics.Google ScholarGoogle Scholar
  10. Funes Mora, K., and Odobez, J. 2014. Geometric generative gaze estimation (g3e) for remote rgb-d cameras. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Guestrin, E. D., and Eizenman, M. 2006. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Transcations on Biomedical Engineering.Google ScholarGoogle Scholar
  12. Guestrin, E. D., and Eizenman, M. 2008. Remote point-of-gaze estimation requiring a single-point calibration for applications with infants. Proceedings of the 2008 symposium on Eye tracking research and applications. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Hansen, D. W., and Ji, Q. 2010. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Transactions on Pattern Analysis and Machine Intelligence. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Harel, J., Koch, C., and Perona, P. 2006. Graph-based visual saliency. NIPS.Google ScholarGoogle Scholar
  15. Harel, J., Koch, C., and Perona, P. 2006. Graph-based visual saliency. In Advances in neural information processing systems, 545--552.Google ScholarGoogle Scholar
  16. Hou, X., and Zhang, L. 2007. Saliency detection: A spectral residual approach. In Computer Vision and Pattern Recognition, 2007. CVPR'07. IEEE Conference on.Google ScholarGoogle Scholar
  17. Itti, L., and Koch, C. 2000. A saliency-based search mechanism for overt and covert shifts of visual attention. Vision research.Google ScholarGoogle Scholar
  18. Itti, L., Koch, C., and Niebur, E. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on pattern analysis and machine intelligence 20, 11, 1254--1259. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Judd, T., Ehinger, K., Durand, F., and Torralba, A. 2009. Learning to predict where humans look. IEEE 12th International Conference on Computer Vision.Google ScholarGoogle Scholar
  20. Judd, T., Durand, F., and Torralba, A. 2012. A benchmark of computational models of saliency to predict human fixations. In MIT Technical Report.Google ScholarGoogle Scholar
  21. Kienzle, W., Franz, M. O., Schölkopf, B., and Wichmann, F. A. 2009. Center-surround patterns emerge as optimal predictors for human saccade targets. Journal of Vision 9, 5, 7.Google ScholarGoogle ScholarCross RefCross Ref
  22. Krizhevsky, A., Sutskever, I., and Hinton, G. E. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, 1097--1105.Google ScholarGoogle Scholar
  23. Kullback, S., and Leibler, R. A. 1951. On information and sufficiency. Ann. Math. Statist. 22, 1 (03), 79--86.Google ScholarGoogle Scholar
  24. Kümmerer, M., Theis, L., and Bethge, M. 2014. Deep gaze i: Boosting saliency prediction with feature maps trained on imagenet. arXiv preprint arXiv:1411.1045.Google ScholarGoogle Scholar
  25. Le, Q. V. 2013. Building high-level features using large scale unsupervised learning. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on.Google ScholarGoogle Scholar
  26. LeCun, Y. 2012. Learning invariant feature hierarchies. In Computer vision--ECCV 2012. Workshops and demonstrations, Springer, 496--505. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Lee, H., Grosse, R., Ranganath, R., and Ng, A. Y. 2009. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th Annual International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Lin, Y., Kong, S., Wang, D., and Zhuang, Y. 2014. Saliency detection within a deep convolutional architecture. In Cognitive Computing for Augmented Human Intelligence Workshop, in conduction with AAAI.Google ScholarGoogle Scholar
  29. Lu, F., Sugano, Y., Okabe, T., and Sato, Y. 2011. Inferring human gaze from appearance via adaptive linear regression. In Proc. International Conference on Computer Vision. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Lu, F., Sugano, Y., Okabe, T., and Sato, Y. 2012. Head pose-free appearance-based gaze sensing via eye image synthesis. International Conference on Pattern Recognition.Google ScholarGoogle Scholar
  31. Maio, W., Chen, J., and Ji, Q. 2011. Constraint-based gaze estimationwithout active calibration. In Automatic Face and Gesture Recognition and Workshops (FG 2011), 2011 IEEE International Conference on.Google ScholarGoogle Scholar
  32. Mancas, M., and Pirri, F. adn Pizzoli, M. 2011. From saliency to eye gaze: embodied visual selection for a pan-tilt-based robotic head. LNCS Series from Proceedings of the 7th International Symposium on Visual Computing (ISVC). Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Mason, M., Hood, B., and Macrae, C. 2004. Look into my eyes: Gaze direction and person memory. Memory.Google ScholarGoogle Scholar
  34. Model, D., and Eizenman, M. 2010. An automatic personal calibration procedure for advanced gaze estimation systems. IEEE Transactions on Biomedical Engineering.Google ScholarGoogle ScholarCross RefCross Ref
  35. Morimoto, C. H., and Mimica, M. R. 2005. Eye gaze tracking techniques for interactive applications. Computer Vision and Image Understanding, Special Issue on Eye Detection and Tracking. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Nuthmann, A., and Henderson, J. M. 2010. Object-based attentional selection in scene viewing. Journal of vision.Google ScholarGoogle ScholarCross RefCross Ref
  37. Olshausen, B. A., and Field, D. J. 1997. Sparse coding with an overcomplete basis set: A strategy employed by v1? Vision research.Google ScholarGoogle Scholar
  38. Olshausen, B. A., et al. 1996. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature.Google ScholarGoogle Scholar
  39. Riche, N., Duvinage, M., Mancas, M., Gosselin, B., and Dutoit, T. 2013. Saliency and human fixations: State-of-the-art and study of comparison metrics. In Computer Vision (ICCV), 2013 IEEE International Conference on. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Riesenhuber, M., and Poggio, T. 1999. Hierarchical models of object recognition in cortex. Nature neuroscience.Google ScholarGoogle Scholar
  41. Rosenholtz, R. 1999. A simple saliency model predicts a number of motion popout phenomena. Vision research.Google ScholarGoogle Scholar
  42. Serre, T., Wolf, L., Bileschi, S., Riesenhuber, M., and Poggio, T. 2007. Robust object recognition with cortex-like mechanisms. Pattern Analysis and Machine Intelligence, IEEE Transactions on. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Shen, C., and Zhao, Q. 2014. Learning to predict eye fixations for semantic contents using multi-layer sparse network. Neurocomputing.Google ScholarGoogle Scholar
  44. Shih, S. W., and Liu, J. 2004. A novel approach to 3-d gaze tracking using stereo cameras. IEEE Transactions on Systems, Man and Cybernetics, PartB. Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Sugano, Y., Matsushita, Y., Sato, Y., and Koike, H. 2007. An incremental learning method for unconstrained gaze estimation. In in European Conference on Computer Vision. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Sugano, Y., Matsushita, Y., and Sato, Y. 2010. Calibration-free gaze sensing using saliency maps. IEEE Conference on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle Scholar
  47. Sugano, Y., Matsushita, Y., and Sato, Y. 2013. Appearance-based gaze estimation using visual saliency. IEEE Trans. Pattern Analysis and Machine Intelligence. Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. Sugano, Y., Matsushita, Y., and Sato, Y. 2014. Learning-by-synthesis for appearance-based 3d gaze estimation. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. Tan, K. H., Kriegman, D., and Ahuja, N. 2002. Appearance-based eye gaze estimation. In Proc. 6th IEEE Workshop on Applications of Computer Vision. Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. Tobii technology. http://www.tobii.com/en/eye-experience/.Google ScholarGoogle Scholar
  51. Treisman, A. M., and Gelade, G. 1980. A feature-integration theory of attention. Cognitive psychology 12, 1, 97--136.Google ScholarGoogle Scholar
  52. Valenti, R., Sebe, N., and Gevers, T. 2011. What are you looking at? improving visual gaze estimation by saliency. Int J of Computer Vision. Google ScholarGoogle ScholarDigital LibraryDigital Library
  53. Vig, E., Dorr, M., and Cox, D. 2014. Large-scale optimization of hierarchical features for saliency prediction in natural images. In Computer Vision and Pattern Recognition, 2014. CVPR'14. IEEE Conference on. Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. Wen, S., Han, J., Zhang, D., and Guo, L. 2014. Saliency detection based on feature learning using deep boltzmann machines. In ICME, 1--6.Google ScholarGoogle Scholar
  55. Williams, O., Blake, A., and Cipolla, R. 2006. Sparse and semi-supervised visual mapping with the s3gp. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. Xu, J., Jiang, M., Wang, S., Kankanhalli, M. S., and Zhao, Q. 2014. Predicting human gaze beyond pixels. Journal of vision.Google ScholarGoogle ScholarCross RefCross Ref
  57. Yamins, D. L., Hong, H., Cadieu, C. F., Solomon, E. A., Seibert, D., and DiCarlo, J. J. 2014. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proceedings of the National Academy of Sciences, 201403112.Google ScholarGoogle ScholarCross RefCross Ref
  58. Zhao, Q., and Koch, C. 2011. Learning a saliency map using fixated locations in natural scenes. Journal of vision.Google ScholarGoogle ScholarCross RefCross Ref
  59. Zhu, Z., and Ji, Q. 2004. Eye and gaze tracking for interactive graphic display. Machine Vision and Applications. Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. Zhu, Z., Ji, Q., and Bennett, K. P. 2006. Nonlinear eye gaze mapping function estimation via support vector regression. International Conference on Pattern Recognition. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Deep eye fixation map learning for calibration-free eye gaze tracking

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        ETRA '16: Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications
        March 2016
        378 pages
        ISBN:9781450341257
        DOI:10.1145/2857491

        Copyright © 2016 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 14 March 2016

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate69of137submissions,50%

        Upcoming Conference

        ETRA '24
        The 2024 Symposium on Eye Tracking Research and Applications
        June 4 - 7, 2024
        Glasgow , United Kingdom

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader