Abstract
Real-time computer vision combined with robust gesture recognition provides a natural alternative to traditional computer interfaces. Human users have plenty of experience with actions and the manipulation of objects requiring finger movement. In place of a mouse, users could use their hands to select and manipulate data. This paper presents a first step in this approach using a finger as a pointing and selection device.
A major feature of a successful tracking system is robustness. The system must be able to acquire tracked features upon startup, and reacquire them if lost during tracking. Reacquisition should be fast and accurate (i.e. it should pick up the correct feature). Intelligent search algorithms are needed for speedy, accurate acquisition of lost features with the frame. The prototype interface presented in this paper is based on finger tracking as a means of input to applications. The focus of the discussion is how the system can be made to perform robustly in real-time. Dynamically distributed search windows are defined for searching within the frame. The location and number of search windows are dependent on the confidence in the tracking of features. Experimental results showing the effectiveness of these techniques are presented.
Preview
Unable to display preview. Download preview PDF.
Bibliography
C. Cadoz, “Les realites virtuelles”, Dominos, Flammarion, 1994.
J. Cassell. “What You Need to Know about Spontaneous Gesture, and Why You Need to Know It”. 2nd International Conference on Automatic Face and Gesture Recognition, Vermont, October 1996.
J. Crowley and J. Coutaz, “Vision for Man Machine Interaction”, EHCI, Grand Targhee, August 1995.
J. Crowley, F. Berard, and J. Coutaz, “Finger Tracking as an Input Device for Augmented Reality”, International Workshop on Gesture and Face Recognition, Zurich, June 1995.
H. Inoue, T. Tachikawa and M. Inaba, “Robot Vision System with a Correlation Chip for Real-time Tracking, Optical Flow and Depth Map Generation”, Proceedings 1992 IEEE International Conference on Robotics and Automation, pp. 1621–1626, 1992.
V. Pavlovic, R. Sharma and T. Huang, “Gestural Interface to a Visual Computing Environment for Molecular Biologists”, Proceedings of the 2d International Conference on Automatic Face and Gesture Recognition, pp. 30–35, 1996.
T. Poston and L. Serra, “The Virtual Workbench: Dextrous VR”, Proceedings ACM VRST'94 — Virtual Reality Software and Technology, pp. 111–122, 1994.
Gesture-driven interaction as a human factor in virtual environments-an approach with neural networks. In R.A. Earnshaw, M.A. Gigante & H. Jones(Eds.), Virtual Reality Systems. London: Academic Press Ltd.
A. Zelinsky and J. Heinzmann, “Real-Time Visual Recognition of Facial Gestures for Human-Computer Interaction”, Proceedings of the 2nd International Conference on Automatic Face and Gesture Recognition, pp. 351–356, 1996.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
O'Hagan, R., Zelinsky, A. (1997). Finger track — A robust and real-time gesture interface. In: Sattar, A. (eds) Advanced Topics in Artificial Intelligence. AI 1997. Lecture Notes in Computer Science, vol 1342. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-63797-4_101
Download citation
DOI: https://doi.org/10.1007/3-540-63797-4_101
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-63797-4
Online ISBN: 978-3-540-69649-0
eBook Packages: Springer Book Archive