ABSTRACT
In this paper, we investigate how head movements can serve to change the viewpoint in 3D applications, especially when the viewpoint needs to be changed quickly and temporarily to disambiguate the view. We study how to use yaw and roll head movements to perform orbital camera control, i.e., to rotate the camera around a specific point in the scene. We report on four user studies. Study 1 evaluates the useful resolution of head movements. Study 2 informs about visual and physical comfort. Study 3 compares two interaction techniques, designed by taking into account the results of the two previous studies. Results show that head roll is more efficient than head yaw for orbital camera control when interacting with a screen. Finally, Study 4 compares head roll with a standard technique relying on the mouse and the keyboard. Moreover, users were allowed to use both techniques at their convenience in a second stage. Results show that users prefer and are faster (14.5%) with the head control technique.
Supplemental Material
- Jonathan Aceituno, Géry Casiez, and Nicolas Roussel. 2013. How Low Can You Go?: Human Limits in Small Unidirectional Mouse Movements. In ACM CHI'13. 1383--1386. Google ScholarDigital Library
- Kevin W. Arthur, Kellogg S. Booth, and Colin Ware. 1993. Evaluating 3D Task Performance for Fish Tank Virtual Worlds. ACM TOIS 11, 3 (1993), 239--265. Google ScholarDigital Library
- Ravin Balakrishnan and Gordon Kurtenbach. 1999. Exploring Bimanual Camera Control and Object Manipulation in 3D Graphics Interfaces (CHI '99). 56--62. Google ScholarDigital Library
- François Bérard. 2014. The Perceptual Window: Head Motion as a new Input Stream. INTERACT (2014), 1--8.Google Scholar
- François Bérard, Guangyu Wang, and Jeremy R. Cooperstock. 2011. On the Limits of the Human Motor Control Precision: The Search for a Device's Human Resolution. In INTERACT'11. Springer-Verlag, 107--122. Google ScholarDigital Library
- Doug A. Bowman, Ernst Kruijff, Joseph J. LaViola, and Ivan Poupyrev. 2004. 3D User Interfaces: Theory and Practice. Addison Wesley Longman Publishing Co., Inc. Google ScholarDigital Library
- BSi. 2003. Visual acuity test types - Part 1: Test charts for clinical determination of distance visual acuity - Specification. BS 4274--1:2003. British Standards Institution.Google Scholar
- Géry Casiez and Daniel Vogel. 2008. The Effect of Spring Stiffness and Control Gain with an Elastic Rate Control Pointing Device (ACM CHI '08). 1709--1718. Google ScholarDigital Library
- Marc Christie, Patrick Olivier, and Jean-Marie Normand. 2008. Camera Control in Computer Graphics. Computer Graphics Forum 27 (2008), 2197--2218.Google ScholarCross Ref
- Andrew Crossan, Mark McGill, Stephen Brewster, and Roderick Murray-Smith. 2009. Head tilting for interaction in mobile contexts. In ACM MobileHCI'09. Google ScholarDigital Library
- Carolina Cruz-Neira, Daniel J. Sandin, and Thomas A. DeFanti. 1993. Surround-screen Projection-based Virtual Reality: The Design and Implementation of the CAVE. In SIGGRAPH '93. ACM, 135--142. Google ScholarDigital Library
- Lawrence D. Cutler, Bernd Fröhlich, and Pat Hanrahan. 1997. Two-handed Direct Manipulation on the Responsive Workbench. In ACM I3D '97. 107--114. Google ScholarDigital Library
- K M Fairchild, B H Lee, J Loo, H Ng, and L Serra. 1993. The heaven and earth virtual reality: Designing applications for novice users. IEEE VR (1993), 47--53. Google ScholarDigital Library
- M. Fiorentino, A. E. Uva, M. Dellisanti Fabiano, and G. Monno. 2010. Improving Bi-manual 3D Input in CAD Modelling by Part Rotation Optimisation. Comput. Aided Des. 42, 5 (May 2010), 462--470. Google ScholarDigital Library
- Paul M. Fitts. 1954. The Information Capacity of the Human Motor System in Controlling the Amplitude of Movement. Journal of Experimental Psychology 47, 6 (1954), 381.Google ScholarCross Ref
- Scott Frees, G Drew Kessler, and Edwin Kay. 2007. PRISM interaction for enhancing control in immersive virtual environments. TOCHI 14, 1 (2007). Google ScholarDigital Library
- Anton Fuhrmann, Dieter Schmalstieg, and Michael Gervautz. 1998. Strolling Through Cyberspace With Your Hands In Your Pockets: Head Directed Navigation In Virtual Environments.Google Scholar
- 1Eleanor J. Gibson, James J. Gibson, Olin W. Smith, and Howard Flock. 1959. Motion parallax as a determinant of perceived depth. Journal of Experimental Psychology (1959), 40--51.Google Scholar
- Thomas Riisgaard Hansen, Eva Eriksson, and Andreas Lykke-Olesen. 2006. Use Your Head: Exploring Face Tracking for Mobile Interaction. In ACM CHI EA'06. 845--850. Google ScholarDigital Library
- Chris Harrison and Anind K. Dey. 2008. Lean and Zoom: Proximity-aware User Interface and Content Magnification. In ACM CHI'08. 507--510. Google ScholarDigital Library
- R. Jagacinski and D. Monk. 1985. Fitts' law in two dimensions with hand and head movements. Journal of Motor Behavior 7 (1985), 77--95.Google ScholarCross Ref
- David R Koller, Mark R Mine, and Scott E Hudson. 1996. Head-tracked orbital viewing: an interaction technique for immersive virtual environments. In ACM UIST '96. 81--82. Google ScholarDigital Library
- Arun Kulshreshth, Joseph J LaViola, and Jr. 2013. Evaluating performance benefits of head tracking in modern video games. In SUI '13. ACM, 53--60. Google ScholarDigital Library
- Joseph Laviola and Michael Katzourin. 2007. An Exploration of Non-Isomorphic 3D Rotation in Surround Screen Virtual Environments. In IEEE 3DUI.Google Scholar
- Edmund LoPresti, David M Brienza, Jennifer Angelo, Lars Gilbertson, and Jonathan Sakai. 2000. Neck range of motion and use of computer head controls. In ACM Assets '00. 121--128. Google ScholarDigital Library
- W.R. Miles. 1930. Ocular Dominance in Human Adults. Journal of General Psychology 3 (1930), 412--429.Google ScholarCross Ref
- Jurriaan D. Mulder and Robert Van Liere. 2000. Enhancing Fish Tank VR. In VR '00. IEEE Computer Society, 91--. Google ScholarDigital Library
- Michaël Ortega and Thomas Vincent. 2014. Direct Drawing on 3D Shapes with Automated Camera Control. In ACM CHI'14. 2047--2050. Google ScholarDigital Library
- Corey Pittman, Joseph J LaViola, and Jr. 2014. Exploring head tracked head mounted displays for first person robot teleoperation. In ACM IUI'14. 323--328. Google ScholarDigital Library
- Ivan Poupyrev, Suzanne Weghorst, and Sidney Fels. 2000. Non-isomorphic 3D rotational techniques. In ACM CHI'00. 540--547. Google ScholarDigital Library
- Ivan Poupyrev, Suzanne Weghorst, Takahiro Otsuka, and Tadao Ichikawa. 1999. Amplifying spatial rotations in 3D interfaces. (1999).Google Scholar
- Wen Qi, Russell M. Taylor, II, Christopher G. Healey, and Jean-Bernard Martens. 2006. A Comparison of Immersive HMD, Fish Tank VR and Fish Tank with Haptics Displays for Volume Visualization. In APGV '06. ACM, 51--58. Google ScholarDigital Library
- R Reilink, G de Bruin, M Franken, M A Mariani, S Misra, and S Stramigioli. 2010. Endoscopic camera control by head movements for thoracic surgery. In BioRob. IEEE, 510--515.Google Scholar
- Adalberto L Simeone, Eduardo Velloso, Jason Alexander, and Hans Gellersen. 2014. Feet movement in desktop 3D interaction. In IEEE 3DUI. 71--74.Google Scholar
- Martin Spindler, Wolfgang Büschel, and Raimund Dachselt. 2012. Use Your Head: Tangible Windows for 3D Information Spaces in a Tabletop Environment. In ACM ITS '12. 245--254. Google ScholarDigital Library
- Robert J. Teather and Wolfgang Stuerzlinger. 2008. Exaggerated Head Motions for Game Viewpoint Control. In Future Play '08. ACM, 240--243. Google ScholarDigital Library
- G.A. Thibodeau and K.T. Patton. 1996. Anatomy and Physiology Third Edition. Mosby.Google Scholar
- Colin Ware, Kevin Arthur, and Kellogg S. Booth. 1993. Fish Tank Virtual Reality. In INTERACT and CHI '93. 37--42. Google ScholarDigital Library
- Kota Yamaguchi, Takashi Komuro, and Masatoshi Ishikawa. 2009. Ptz control with head tracking for video chat. In CHI EA '09. 3919--3924. Google ScholarDigital Library
Index Terms
- Desktop Orbital Camera Motions Using Rotational Head Movements
Recommendations
Eye&Head: Synergetic Eye and Head Movement for Gaze Pointing and Selection
UIST '19: Proceedings of the 32nd Annual ACM Symposium on User Interface Software and TechnologyEye gaze involves the coordination of eye and head movement to acquire gaze targets, but existing approaches to gaze pointing are based on eye-tracking in abstraction from head motion. We propose to leverage the synergetic movement of eye and head, and ...
Eye gaze tracking with free head movements using a single camera
SoICT '10: Proceedings of the 1st Symposium on Information and Communication TechnologyThe problem of eye gaze tracking has been researched and developed for a long time. The most difficult problem in the non-intrusive system of eye gaze tracking is the problem of head movements. Some of existing methods have to use two cameras and an ...
Head or gaze?: controlling remote camera for hands-busy tasks in teleoperation: a comparison
OZCHI '10: Proceedings of the 22nd Conference of the Computer-Human Interaction Special Interest Group of Australia on Computer-Human InteractionHead motion and eye gaze are general models of natural human interaction. Recent computer vision based head tracking and eye tracking technologies have expanded the possibilities of designing and developing more natural and intuitive user interfaces for ...
Comments