Abstract
We designed and implemented a vision-based projected tabletop interface for finger interaction. The system offers a simple and quick setup and economic design. The projection onto the tabletop provides more comfortable and direct viewing for users, and more natural, intuitive yet flexible interaction than classical or tangible interfaces. Homography calibration techniques are used to provide geometrically compensated projections on the tabletop. A robust finger tracking algorithm is proposed to enable accurate and efficient interactions using this interface. Two applications have been implemented based on this interface.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Sukthankar, R., Stockton, R., Mullin, M.: Smarter Presentations: Exploiting homography in camera-projector systems. In: Proc. International Conference on Computer Vision, Vancouver, Canada, pp. 247–253 (2001)
Chen, H., Wallace, G., Gupta, A., Li, K., Funkhouser, T., Cook, P.: Experiences with scalability of display walls. In: Proc. Immersive Projection Technology Symposium (IPT), Orlando, FL (2002)
Czernuszenko, M., Pape, D., Sandin, D., DeFanti, T., Dawe, L., Brown, M.: The ImmersaDesk and InfinityWall projection-based virtual reality displays. Computer Graphics, 46–49 (1997)
Buxton, W., Fitzmaurice, G., Balakrishnan, R., Kurtenbach, G.: Large displays in automotive design. IEEE Computer Graphics and Applications 20(4), 68–75 (2000)
Song, P., Winkler, S., Tedjokusumo, J.: A Tangible Game Interface Using Projector-Camera Systems. In: HCI 2007. LNCS, vol. 4551, pp. 956–965. Springer, Heidelberg (2007)
Patten, J., Ishii, H., Pangaro, G.: Sensetable: A wireless object tracking platform for tangible user interfaces. In: Proc. CHI, Conference on Human Factors in Computing Systems, Seattle, Washington, USA (2001)
Mynatt, E.D., Igarashi, T., Edwards, W.K.: Flatland: New dimensions in office whiteboards. In: Proc. CHI 1999, Pittsburgh, PA, USA (1999)
Ashdown, M., Robinson, P.: Escritoire: A personal projected display. IEEE Multimedia Magazine 12(1), 34–42 (2005)
Leigh, D., Dietz, P.: DiamondTouch characteristics and capabilities. In: UbiComp 2002 Workshop on Collaboration with Interactive Tables and Walls, Göteborg, Sweden (2002)
Rekimoto, J.: SmartSkin: An infrastructure for freehand manipulation on interactive surfaces. In: Proc. CHI 2002, Göteborg, Sweden, pp. 113–120 (2002)
Sukthankar, R., Stockton, R., Mullin, M.: Automatic keystone correction for camera-assisted presentation interfaces. In: Tan, T., Shi, Y., Gao, W. (eds.) ICMI 2000. LNCS, vol. 1948, pp. 607–614. Springer, Heidelberg (2000)
Lien, C., Huang, C.: Model-based articulated hand motion tracking for gesture recognition. Image and Vision Computing 16(2), 121–134 (1998)
Triesch, J., Malsburg, C.: Robust classification of hand postures against complex background. In: Proc. International Conference On Automatic Face and Gesture Recognition, Killington (1996)
Segen, J.: Gesture VR: Vision-based 3D hand interface for spatial interaction. In: Proc. ACM Multimedia Conference, Bristol, UK, ACM Press, New York (1998)
Rehg, J., Kanade, T.: Digiteyes: Vision-based 3D human hand tracking. In: Technical Report CMU-CS-93-220, School of Computer Science, Carnegie Mellon University (1993)
Sato, Y., Kobayashi, Y., Koike, H.: Fast tracking of hands and fingertips in infrared images for augmented desk interface. In: Proc. International Conference on Automatic Face and Gesture Recognition, Grenoble, France (2000)
Crowley, J., BĂ©rard, F., Coutaz, J.: Finger tracking as an input device for augmented reality. In: Proc. International Conference on Automatic Face and Gesture Recognition, ZĂĽrich, Switzerland (1995)
Laptev, I., Lindeberg, T.: Tracking of multi-state hand models using particle filtering and a hierarchy of multi-scale image features. Technical Report ISRN KTH/NA/P-00/12-SE, The Royal Institute of Technology (KTH), Stockholm, Sweden (2000)
Hardenberg, C., Brard, F.: Bare-hand human computer interaction. In: Proc. Perceptual User Interfaces, Orlando, Florida, USA (2001)
Microsoft Corporation: Microsoft Windows XP Tablet PC Edition 2005 Recognizer Pack, http://www.microsoft.com/downloads/details.aspx?familyid=080184dd-5e92-4464-b907-10762e9f918b&displaylang=en
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Song, P., Winkler, S., Gilani, S.O., Zhou, Z. (2007). Vision-Based Projected Tabletop Interface for Finger Interactions. In: Lew, M., Sebe, N., Huang, T.S., Bakker, E.M. (eds) Human–Computer Interaction. HCI 2007. Lecture Notes in Computer Science, vol 4796. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-75773-3_6
Download citation
DOI: https://doi.org/10.1007/978-3-540-75773-3_6
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-75772-6
Online ISBN: 978-3-540-75773-3
eBook Packages: Computer ScienceComputer Science (R0)