Abstract
In this paper we present three concepts that address movement-based interaction using camera tracking. Based on our work with several movement-based projects we present four selected applications, and use these applications to leverage our discussion, and to describe our three main concepts space, relations, and feedback. We see these as central for describing and analysing movement-based systems using camera tracking and we show how these three concepts can be used to analyse other camera tracking applications.
Similar content being viewed by others
References
Barkhuus M, Chalmers M, Tennent, P, Bell M, Hall M, Sherwood S, Brown B (2005) Picking pockets on the lawn: the development of tactics and strategies in a mobile game. In: Proceedings of ubicomp, Springer 3-540-28760-4, pp 358–374
Dance Dance Revolution UltraMix 3. Available at http://www.konami-usa.com/production/ddr_ultramix3/index.html
Samsung SCH-S310Available at http://www.samsung.com/
Patridge K, Chatterjee S, Sazawal V, Borriello G, Want R (2002) TiltType: accelerometer-supported text entry for very small devices. In: Proceedings of the 15th annual ACM symposium on user interface software and technology UIST’02, Paris, 27–30 October 2002, pp 201–204
Pavlovic V, Sharma R, Huang T (1997) Visual Interpretation of Hand Gestures for Human-Computer Interaction. IEEE Trans Pattern Anal Machine Intell 19(7):677–695
Bradski G, (1998) Computer vision face tracking for use in a perceptual user interface, intel technology. J Q2, pp 1–15
Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction space – designing for camera based interaction with mobile devices. In: Conference on human factors in computing systems in CHI ’05 extended abstracts on human factors in computing systems, Portland, ACM Press
Mackay W (1998) Augmented reality: linking real and virtual worlds: a new paradigm for interacting with computers. In: Proceedings of the working conference on advanced visual interfaces, L’Aquila, 24–27 May 1998
Benford S, Gaver B, Boucher A, Walker B, Pennington S, Schmidt A, Gellersen H, Steed A., Schnädelbach H, Koleva B, Anastasi R, Greenhalgh C, Rodden T, Green J, Ghali A, Pridmore T (2005) Expected, sensed, and desired: a framework for designing sensing-based interaction. ACM transactions on computer-human interaction (TOCHI) vol. 12 , Issue 1. ACM Press, pp 3–30
Abowd G, Mynatt E (2000) Charting past, present, and future research in ubiquitous computing. In: ACM transactions on computer–human interaction (TOCHI), vol 7 , Issue 1. Special issue on human-computer interaction in the new millennium, part 1, ACM Press, pp 29–58
Ullmer B, Ishii H (2000) Emerging frameworks for tangible user interfaces. IBM Syst J 39(3–4):915–931
Holmquist LE, Redström J, Ljungstrand P (1999) Token-based access to digital information. In: Proceedings of the 1st international symposium on handheld and ubiquitous computing (HUC’99), Karlsruhe, pp 234–245
Fishkin K (2004) A taxonomy for and analysis of tangible interfaces. Pers Ubiquit Comput 8(5):347–358
Underkoffler J, Ishii H (1999) Urp: a luminous-tangible workbench for urban planning and design. In: Conference on human factors in computing systems, proceedings of the SIGCHI conference on human factors in computing systems. Pittsburgh, Pennsylvania, pp 386–393
Krogh PG, Ludvigsen M, Lykke-Olesen A (2004) “Help Me Pull That Cursor”—a collaborative interactive floor enhancing community interaction. In: Proceedings of OZCHI 2004, Wollongong, CD-ROM, 22–24 November 2004. ISBN:1 74128 079
StorySurfer. Available at http://www.interactivespaces.net
Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction spaces—expanding the interaction space with mobile devices. In: Proceedings of the 19th British HCI Group annual conference, Edingburgh, 5–9 September 2005
Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction spaces—a new interaction technique for mobile devices. In: Proceedings of the 7th international conference on ubiquitous computing, Tokyo
Forty A (2000) Words and buildings—a vocabulary of modern architecture. Thames and Hudson, New York
Gehl J, (1987) Life between buildings—using public space, Van Nostrand Reinhold, New York
Bellotti V, Back M, Edwards WK, Henderson A, Lopes C. (2002) Making sense of sensing systems: five questions for designers and researchers. Conference on human factors in computing systems. In: Proceedings of CHI 2002, ACM Press, Minneapolis, pp 471–509
Larssen AT, Loke L, Robertson T, Edwards J (2004) Movement as input for interaction—a study and evaluation of two eyetoy (TM) games. In: Proceedings of OZCHI 2004, Wollongong, CD-ROM, 22–24 November 2004. ISBN:1 74128 079
Bødker S (1996) Applying activity theory to video analysis: how to make sense of video data in human-computer interaction. In: Nardi B (ed) Context and consciousness. Activity theory and human computer interaction, MIT press, pp 147–174
Sony Eye-Toy. Available at http://www.eyetoy.com
Lyons M, Haehnel M, Tetsutani N (2001) The mouthesizer: a facial gesture musical interface. conference abstracts, Siggraph 2001, Los Angeles, p 230
Hämäläinen P, Ilmonen T, Höysniemi J, Lindholm M, Nykänen A (2005) Martial arts in artificial reality. In: Proceedings of CHI 2005 conference on human factors in computing systems, Portland, pp 781–790
Henrysson A, Billinghurst M, Ollila M (2005) Face to face collaborative AR on mobile phones. In: Proceedings of the international symposium on mixed and augmented reality (ISMAR 2005), Vienna
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Eriksson, E., Hansen, T.R. & Lykke-Olesen, A. Movement-based interaction in camera spaces: a conceptual framework. Pers Ubiquit Comput 11, 621–632 (2007). https://doi.org/10.1007/s00779-006-0134-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00779-006-0134-z