Abstract
Many studies on eye tracking have been conducted in diverse research areas. Nevertheless, eye tracking continues to be limited by low accuracy and a severe vibration problem due to pupil tremors. Furthermore, because almost all selection interactions, such as click events, use a dwell-time or eye-blinking method, eye tracking presents issues for both time consumption and involuntary blinking. In this paper, we therefore propose a multi-modal interaction method using a combination of eye tracking and hand gesture recognition with the commercial hand gesture controller. This method performs global and intuitive navigation using eye tracking, and local and detailed navigation using hand gesture controller. It supports intuitive hand gestures for mouse-button clicking. Experimental results indicate that the targeting time for small points is significantly improved using the proposed method. Especially, the proposed method has advantages in large display with high spatial resolution environment. Also, the proposed clicking interaction and modality switching concept showed accurate recognition rate and positive training effect, respectively.
Similar content being viewed by others
References
MacKenzie IS, Zhang X (2008) Eye typing using word and letter prediction and a fixation algorithm. In: 2008 symposium on eye tracking research and applications, pp 55–58
Zhu Z, Ji Q (2004) Eye and gaze tracking for interactive graphic display. Mach Vis Appl 15(July):139–148
Cutrell E, Guan Z (2007) What are you looking for? An eye-tracking study of information usage in web Search. SIGCHI conference on human factors in computing systems, pp 407–416
Pan B, Hembrooke HA, Gay GK, Granka LA, Feusner MK, Newman JK (2004) Determinants of web page viewing behavior: an eye-tracking study. In: 2004 symposium on eye tracking research and applications, pp 147–154
Lee EC, Park KR (2007) A study on eye gaze estimation method based on cornea model of human eye. Comput Vis/Comput Gr Collab Tech 4418(March):307–317
Suh KH, Kim Y, Kim Y, Ko D, Lee EC (2015) Monocular eye tracking system using webcam and zoom lens. Adv Multimed Ubiquitous Eng 352(June):135–141
Grauman K, Betke M, Gips J, Bradski GR (2001) Communication via eye blinks: detection and duration analysis in real time. In: 2001 IEEE computer society conference on computer vision and pattern recognition, vol 1, pp I-1010-I-1017
Grauman K, Betke M, Lombardi J, Gips J, Bradski GR (2003) Communication via eye blinks and eyebrow raises: video-based human-computer interfaces. Univers Access Inf Soc 2:2–4
Yamamoto Y, Yoda I, Sakaue K (2004) Arm-pointing gesture interface using surrounded stereo cameras system. In: 17th international conference on pattern recognition, vol 4, pp 965–970
Jing P, Yepeng G (2013) Human-computer Interaction using pointing gesture based on an adaptive virtual touch screen. J Signal Process Pattern Recognit 6(4):81–92
Yepeng G, Mingen Z (2008) Real-time 3D pointing gesture recognition for natural HCI. In: 7th World Congress on intelligent control and automation, pp 2433–2436
Carbini S, Viallet JE, Bernier O (2004) Pointing gesture visual recognition for large display. In: International workshop on visual observation of deictic gestures, pp 27–32
Park H, Choi J, Park J, Moon K (2013) A study on hand region detection for kinect-based hand shape recognition. Korean Soc Broadcast Eng 18(3):393–400
Choi J, Park H, Park J (2011) Hand shape recognition using distance transform and shape decomposition. In: 18th IEEE international conference on image processing (ICIP), pp 3605–3608
Oikonomidis I, Kyriazis N, Argyros A (2010) Markerless and efficient 26-DOF hand pose recovery. In: 10th Asian conference on computer vision, pp 744–757
Cao C, Sun Y, Li R, Chen L (2011) Hand posture recognition via joint feature sparse representation. Optical Eng 50(12):127210
Shin G, Chun J (2007) Vision-based multimodal human computer interface based on parallel tracking of eye and hand motion. In: 2007 international conference on convergence information technology, pp 2443–2448
Bachmann D, Weichert F, Rinkenauer G (2014) Evaluation of the leap motion controller as a new contact-free pointing device. Sensors 15:214–233
http://sourceforge.net/projects/opencvlibrary. Accessed 20 Sept 2015
http://ashlandtech.org/2014/04/02/product-comparison-kinect-and-leap-motion/. Accessed 30 Sept 2015
Guna J, Jakus G, Pogacnik M, Tomazic S, Sodnik J (2014) An analysis of the precision and reliability of the leap motion sensor and its suitability for static and dynamic tracking. Sensors 14:3702–3720
https://developer.leapmotion.com/documentation/cpp/devguide/Leap_Coordinate_Mapping.html. Accessed 30 Sept 2015
http://blog.leapmotion.com/getting-started-leap-motion-sdk/. Accessed 30 Sept 2015
SetCursorPos Function (Windows), http://msdn.microsoft.com/en-us/library/windows/desktop/ms648394(v=vs.85).aspx. Accessed 30 Sept 2015
Acknowledgements
This work was supported by the ICT R&D program of MSIP/IITP. [R0126-15-1045, The development of technology for social life logging based on analysis of social emotion and intelligence of convergence content].
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Kim, H., Suh, K.H. & Lee, E.C. Multi-modal user interface combining eye tracking and hand gesture recognition. J Multimodal User Interfaces 11, 241–250 (2017). https://doi.org/10.1007/s12193-017-0242-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12193-017-0242-2