ABSTRACT
This research explores the use of eye-tracking during Augmented Reality (AR) - supported conversations. In this scenario, users can obtain information that supports the conversation, without augmentations distracting the actual conversation.We propose using gaze that allows users to gradually reveal information on demand. Information is indicated around user’s head, which becomes fully visible when other’s visual attention explicitly falls upon the area. We describe the design of such an AR UI and present an evaluation of the feasibility of the concept. Results show that despite gaze inaccuracies, users were positive about augmenting their conversations with contextual information and gaze interactivity. We provide insights into the trade-offs between focusing on the task at hand (i.e., the conversation), and consuming AR information. These findings are useful for future use cases of eye based AR interactions by contributing to a better understanding of the intricate balance between informative AR and information overload.
Supplemental Material
- Antti Ajanki, Mark Billinghurst, Hannes Gamper, Toni Järvenpää, Melih Kandemir, Samuel Kaski, Markus Koskela, Mikko Kurimo, Jorma Laaksonen, Kai Puolamäki, 2011. An augmented reality interface to contextual information. Virtual reality 15, 2-3 (2011), 161–173.Google Scholar
- Ronald Azuma, Yohan Baillot, Reinhold Behringer, Steven Feiner, Simon Julier, and Blair MacIntyre. 2001. Recent Advances in Augmented Reality. IEEE Comput. Graph. Appl. 21, 6 (Nov. 2001), 34–47. https://doi.org/10.1109/38.963459Google ScholarDigital Library
- Blaine Bell, Steven Feiner, and Tobias Höllerer. 2001. View Management for Virtual and Augmented Reality(UIST ’01). ACM, New York, NY, USA, 101–110. https://doi.org/10.1145/502348.502363Google Scholar
- Richard A Bolt. 1981. Gaze-orchestrated dynamic windows. In ACM SIGGRAPH Computer Graphics, Vol. 15. ACM, 109–119.Google Scholar
- Rakesh D Desale and Vandana S Ahire. 2013. A study on wearable gestural interface–a sixthsense technology. IOSR Journal of Computer Engineering (IOSR-JCE) e-ISSN (2013), 2278–0661.Google Scholar
- Steven Feiner, Blair MacIntyre, Marcus Haupt, and Eliot Solomon. 1993. Windows on the World: 2D Windows for 3D Augmented Reality(UIST ’93). ACM, New York, NY, USA, 145–155. https://doi.org/10.1145/168642.168657Google Scholar
- Bernardo A Huberman and Fang Wu. 2007. The economics of attention: maximizing user value in information-rich environments. In Proceedings of the 1st international workshop on Data mining and audience intelligence for advertising. ACM, 16–20.Google ScholarDigital Library
- Yoshio Ishiguro and Jun Rekimoto. 2011. Peripheral Vision Annotation: Noninterference Information Presentation Method for Mobile Augmented Reality(AH ’11). ACM, New York, NY, USA, Article 8, 5 pages. https://doi.org/10.1145/1959826.1959834Google Scholar
- Robert JK Jacob. 1993. Eye movement-based human-computer interaction techniques: Toward non-command interfaces. Advances in human-computer interaction 4 (1993), 151–190.Google Scholar
- Robert J. K. Jacob. 1990. What You Look at is What You Get: Eye Movement-based Interaction Techniques(CHI ’90). ACM, New York, NY, USA, 11–18. https://doi.org/10.1145/97243.97246Google Scholar
- Simon Julier, Yohan Baillot, Dennis Brown, and Marco Lanzagorta. 2002. Information Filtering for Mobile Augmented Reality. IEEE Comput. Graph. Appl. 22, 5 (Sept. 2002), 12–15. https://doi.org/10.1109/MCG.2002.1028721Google ScholarDigital Library
- Jens Keil, Michael Zoellner, Timo Engelke, Folker Wientapper, and Michael Schmitt. 2013. Controlling and filtering information density with spatial interaction techniques via handheld augmented reality. In International Conference on Virtual, Augmented and Mixed Reality. Springer, 49–57.Google ScholarCross Ref
- Mirae Kim, Min Kyung Lee, and Laura Dabbish. 2015. Shop-i: Gaze based Interaction in the Physical World for In-Store Social Shopping Experience. In Proceedings of the 33rd Annual ACM Conference Extended Abstracts on Human Factors in Computing Systems(CHI EA ’15). Association for Computing Machinery, Seoul, Republic of Korea, 1253–1258. https://doi.org/10.1145/2702613.2732797Google ScholarDigital Library
- Mikko Kytö, Barrett Ens, Thammathip Piumsomboon, Gun A. Lee, and Mark Billinghurst. 2018. Pinpointing: Precise Head- and Eye-Based Target Selection for Augmented Reality(CHI ’18). ACM, New York, NY, USA, Article 81, 14 pages. https://doi.org/10.1145/3173574.3173655Google ScholarDigital Library
- Ann McNamara, Katherine Boyd, David Oh, Ryan Sharpe, and Annie Suther. 2018. Using Eye Tracking to Improve Information Retrieval in Virtual Reality. In 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct). 242–243. https://doi.org/10.1109/ISMAR-Adjunct.2018.00076 ISSN: null.Google ScholarCross Ref
- Ann McNamara and Chethna Kabeerdoss. 2016. Mobile Augmented Reality: Placing Labels Based on Gaze Position. In 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct). 36–37. https://doi.org/10.1109/ISMAR-Adjunct.2016.0033 ISSN: null.Google Scholar
- Jakob Nielsen. 1993. Noncommand User Interfaces. Commun. ACM 36, 4 (April 1993), 83–99. https://doi.org/10.1145/255950.153582Google ScholarDigital Library
- Anjul Patney, Marco Salvi, Joohwan Kim, Anton Kaplanyan, Chris Wyman, Nir Benty, David Luebke, and Aaron Lefohn. 2016. Towards Foveated Rendering for Gaze-Tracked Virtual Reality. ACM Trans. Graph. 35, 6, Article Article 179 (Nov. 2016), 12 pages. https://doi.org/10.1145/2980179.2980246Google ScholarDigital Library
- Marcus Tönnis and Gudrun Klinker. 2014. Boundary Conditions for Information Visualization with Respect to the User’s Gaze(AH ’14). ACM, New York, NY, USA, Article 44, 8 pages. https://doi.org/10.1145/2582051.2582095Google Scholar
- Roel Vertegaal 2003. Attentive user interfaces. Commun. ACM 46, 3 (2003), 30–33.Google ScholarDigital Library
- Sean White and Steven Feiner. 2009. SiteLens: Situated Visualization Techniques for Urban Site Visits(CHI ’09). ACM, New York, NY, USA, 1117–1120. https://doi.org/10.1145/1518701.1518871Google Scholar
Recommendations
Advantages of eye-gaze over head-gaze-based selection in virtual and augmented reality under varying field of views
COGAIN '18: Proceedings of the Workshop on Communication by Gaze InteractionThe current best practice for hands-free selection using Virtual and Augmented Reality (VR/AR) head-mounted displays is to use head-gaze for aiming and dwell-time or clicking for triggering the selection. There is an observable trend for new VR and AR ...
Exploring Gaze-assisted and Hand-based Region Selection in Augmented Reality
ETRARegion selection is a fundamental task in interactive systems. In 2D user interfaces, users typically use a rectangle selection tool to formulate a region using a mouse or touchpad. Region selection in 3D spaces, especially in Augmented Reality (AR) ...
Looking for Info: Evaluation of Gaze Based Information Retrieval in Augmented Reality
Human-Computer Interaction – INTERACT 2021AbstractThis paper presents the results of an empirical study and a real-world deployment of a gaze-adaptive UI for Augmented Reality (AR). AR introduces an attention dilemma between focusing on the reality vs. on AR content. Past work suggested eye gaze ...
Comments