ABSTRACT
Previous work on eye tracking and eye-based human-computer interfaces mainly concentrated on making use of the eyes in traditional desktop settings. With the recent growth of interest in smart glass devices and low-cost eye trackers, however, gaze-based techniques for mobile computing is becoming increasingly important. PETMEI 2014 focuses on the pervasive eye tracking paradigm as a trailblazer for mobile eye-based interaction and eye-based context-awareness. We want to stimulate and explore the creativity of these communities with respect to the implications, key research challenges, and new applications for pervasive eye tracking in ubiquitous computing. The long-term goal is to create a strong interdisciplinary research community linking these fields together and to establish the workshop as the premier forum for research on pervasive eye tracking.
- Baldauf, M., Fröhlich, P., and Hutter, S. Kibitzer: a wearable system for eye-gaze-based mobile urban exploration. In Proc. AH2010 (2010), 1--5. Google ScholarDigital Library
- Bulling, A., Cheng, S., Brône, G., and Majaranta, P. 2nd international workshop on pervasive eye tracking and mobile eye-based interaction (petmei 2012): Proposal for a workshop (mini-track) at ubicomp 2012. In Proc. UbiComp '12 (2012), 673--676. Google ScholarDigital Library
- Bulling, A., Duchowski, A. T., and Majaranta, P. Petmei 2011: The 1st international workshop on pervasive eye tracking and mobile eye-based interaction. In Proc. UbiComp '11 (2011), 627--628. Google ScholarDigital Library
- Bulling, A., Ward, J. A., Gellersen, H., and Tröster, G. Eye Movement Analysis for Activity Recognition Using Electrooculography. IEEE Trans. on Pattern Analysis and Machine Intelligence 33, 4 (2011), 741--753. Google ScholarDigital Library
- Duchowski, A. T. Eye Tracking Methodology: Theory and Practice. Springer, 2007. Google ScholarDigital Library
- Fathi, A., Li, Y., and Rehg, J. M. Learning to recognize daily actions using gaze. In Proc. ECCV 2012. Springer, 2012, 314--327. Google ScholarDigital Library
- Foulsham, T., and Kingstone, A. Goal-driven and bottom-up gaze in an active real-world search task. In Proc. ETRA '12 (2012), 189--192. Google ScholarDigital Library
- Gergle, D., and Clark, A. See what i'm saying?: using dyadic mobile eye tracking to study collaborative reference. In Proc. CSCW 2011, ACM (2011), 435--444. Google ScholarDigital Library
- Hansen, D. W., and Ji, Q. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Trans. on Pattern Analysis and Machine Intelligence 32, 3 (2010), 478--500. Google ScholarDigital Library
- Ishiguro, Y., Mujibiya, A., Miyaki, T., and Rekimoto, J. Aided eyes: Eye activity sensing for daily life. In Proc. AH2010 (2010). Google ScholarDigital Library
- Karthikeyan, S., Jagadeesh, V., Shenoy, R., Ecksteinz, M., and Manjunath, B. From where and how to what we see. In Proc. ICCV2013 (2013). Google ScholarDigital Library
- Komogortsev, O. V., Jayarathna, S., Koh, D. H., and Gowda, S. M. Qualitative and quantitative scoring and evaluation of the eye movement classification algorithms. In Proc. ETRA 2010, ACM Press (2010), 65--68. Google ScholarDigital Library
- Kunze, K., Ishimaru, S., Utsumi, Y., and Kise, K. My reading life: Towards utilizing eyetracking on unmodified tablets and phones. In Proc. UbiComp '13 Adjunct (2013), 283--286. Google ScholarDigital Library
- Ogaki, K., Kitani, K., Sugano, Y., and Sato, Y. Coupling eye-motion and ego-motion features for first-person activity recognition. In Second IEEE Workshop on Egocentric Vision (CVPRW) (2012), 1--7.Google ScholarCross Ref
- Park, H. S., Jain, E., and Sheikh, Y. 3d social saliency from head-mounted cameras. In Proc. NIPS 2012 (2012), 431--439.Google Scholar
- Pfeiffer, T., and Renner, P. Eyesee3d: A low-cost approach for analyzing mobile 3d eye tracking data using computer vision and augmented reality technology. In Proc. ETRA '14 (2014), 195--202. Google ScholarDigital Library
- Salvucci, D. D., and Anderson, J. R. Automated eye-movement protocol analysis. Human-Computer Interaction 16, 1 (2001), 39--86. Google ScholarDigital Library
- Toyama, T., Kieninger, T., Shafait, F., and Dengel, A. Gaze guided object recognition using a head-mounted eye tracker. In Proc. ETRA '12 (2012), 91--98. Google ScholarDigital Library
- Wood, E., and Bulling, A. Eyetab: Model-based gaze estimation on unmodified tablet computers. In Proc. ETRA '14 (2014), 207--210. Google ScholarDigital Library
- Yun, K., Peng, Y., Samaras, D., Zelinsky, G. J., and Berg, T. L. Studying relationships between human gaze, description, and computer vision. In Proc. CVPR 2013 (2013), 739--746. Google ScholarDigital Library
Index Terms
- 4th international workshop on pervasive eye tracking and mobile eye-based interaction
Recommendations
Eye Tracking in Human-Computer Interaction and Usability Research
INTERACT '09: Proceedings of the 12th IFIP TC 13 International Conference on Human-Computer Interaction: Part IIThe objective of the tutorial is to give an overview on how eye tracking is currently used and how it can be used as a method in human computer interaction research and especially in usability research. An eye tracking system records how the eyes move ...
6th international workshop on pervasive eye tracking and mobile eye-based interaction
UbiComp '16: Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: AdjunctPrevious work on eye tracking and eye-based human-computer interfaces mainly concentrated on making use of the eyes in traditional desktop settings. With the recent growth of interest in wearable computers, such as smartwatches, smart eyewears and low-...
The 5th international workshop on pervasive eye tracking and mobile eye-based interaction
UbiComp/ISWC'15 Adjunct: Adjunct Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2015 ACM International Symposium on Wearable ComputersPrevious work on eye tracking and eye-based human-computer interfaces mainly concentrated on making use of the eyes in traditional desktop settings. With the recent growth of interest in smart eyewear and low-cost mobile eye trackers, gaze-based ...
Comments