Skip to main content
Log in

Movement-based interaction in camera spaces: a conceptual framework

  • Original Article
  • Published:
Personal and Ubiquitous Computing Aims and scope Submit manuscript

Abstract

In this paper we present three concepts that address movement-based interaction using camera tracking. Based on our work with several movement-based projects we present four selected applications, and use these applications to leverage our discussion, and to describe our three main concepts space, relations, and feedback. We see these as central for describing and analysing movement-based systems using camera tracking and we show how these three concepts can be used to analyse other camera tracking applications.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Barkhuus M, Chalmers M, Tennent, P, Bell M, Hall M, Sherwood S, Brown B (2005) Picking pockets on the lawn: the development of tactics and strategies in a mobile game. In: Proceedings of ubicomp, Springer 3-540-28760-4, pp 358–374

  2. Dance Dance Revolution UltraMix 3. Available at http://www.konami-usa.com/production/ddr_ultramix3/index.html

  3. Samsung SCH-S310Available at http://www.samsung.com/

  4. Patridge K, Chatterjee S, Sazawal V, Borriello G, Want R (2002) TiltType: accelerometer-supported text entry for very small devices. In: Proceedings of the 15th annual ACM symposium on user interface software and technology UIST’02, Paris, 27–30 October 2002, pp 201–204

  5. Pavlovic V, Sharma R, Huang T (1997) Visual Interpretation of Hand Gestures for Human-Computer Interaction. IEEE Trans Pattern Anal Machine Intell 19(7):677–695

    Article  Google Scholar 

  6. Bradski G, (1998) Computer vision face tracking for use in a perceptual user interface, intel technology. J Q2, pp 1–15

  7. Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction space – designing for camera based interaction with mobile devices. In: Conference on human factors in computing systems in CHI ’05 extended abstracts on human factors in computing systems, Portland, ACM Press

  8. Mackay W (1998) Augmented reality: linking real and virtual worlds: a new paradigm for interacting with computers. In: Proceedings of the working conference on advanced visual interfaces, L’Aquila, 24–27 May 1998

  9. Benford S, Gaver B, Boucher A, Walker B, Pennington S, Schmidt A, Gellersen H, Steed A., Schnädelbach H, Koleva B, Anastasi R, Greenhalgh C, Rodden T, Green J, Ghali A, Pridmore T (2005) Expected, sensed, and desired: a framework for designing sensing-based interaction. ACM transactions on computer-human interaction (TOCHI) vol. 12 , Issue 1. ACM Press, pp 3–30

  10. Abowd G, Mynatt E (2000) Charting past, present, and future research in ubiquitous computing. In: ACM transactions on computer–human interaction (TOCHI), vol 7 , Issue 1. Special issue on human-computer interaction in the new millennium, part 1, ACM Press, pp 29–58

  11. Ullmer B, Ishii H (2000) Emerging frameworks for tangible user interfaces. IBM Syst J 39(3–4):915–931

    Article  Google Scholar 

  12. Holmquist LE, Redström J, Ljungstrand P (1999) Token-based access to digital information. In: Proceedings of the 1st international symposium on handheld and ubiquitous computing (HUC’99), Karlsruhe, pp 234–245

  13. Fishkin K (2004) A taxonomy for and analysis of tangible interfaces. Pers Ubiquit Comput 8(5):347–358

    Article  Google Scholar 

  14. Underkoffler J, Ishii H (1999) Urp: a luminous-tangible workbench for urban planning and design. In: Conference on human factors in computing systems, proceedings of the SIGCHI conference on human factors in computing systems. Pittsburgh, Pennsylvania, pp 386–393

  15. Krogh PG, Ludvigsen M, Lykke-Olesen A (2004) “Help Me Pull That Cursor”—a collaborative interactive floor enhancing community interaction. In: Proceedings of OZCHI 2004, Wollongong, CD-ROM, 22–24 November 2004. ISBN:1 74128 079

  16. StorySurfer. Available at http://www.interactivespaces.net

  17. Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction spaces—expanding the interaction space with mobile devices. In: Proceedings of the 19th British HCI Group annual conference, Edingburgh, 5–9 September 2005

  18. Hansen TR, Eriksson E, Lykke-Olesen A (2005) Mixed interaction spaces—a new interaction technique for mobile devices. In: Proceedings of the 7th international conference on ubiquitous computing, Tokyo

  19. Forty A (2000) Words and buildings—a vocabulary of modern architecture. Thames and Hudson, New York

    Google Scholar 

  20. Gehl J, (1987) Life between buildings—using public space, Van Nostrand Reinhold, New York

  21. Bellotti V, Back M, Edwards WK, Henderson A, Lopes C. (2002) Making sense of sensing systems: five questions for designers and researchers. Conference on human factors in computing systems. In: Proceedings of CHI 2002, ACM Press, Minneapolis, pp 471–509

  22. Larssen AT, Loke L, Robertson T, Edwards J (2004) Movement as input for interaction—a study and evaluation of two eyetoy (TM) games. In: Proceedings of OZCHI 2004, Wollongong, CD-ROM, 22–24 November 2004. ISBN:1 74128 079

  23. Bødker S (1996) Applying activity theory to video analysis: how to make sense of video data in human-computer interaction. In: Nardi B (ed) Context and consciousness. Activity theory and human computer interaction, MIT press, pp 147–174

  24. Sony Eye-Toy. Available at http://www.eyetoy.com

  25. Lyons M, Haehnel M, Tetsutani N (2001) The mouthesizer: a facial gesture musical interface. conference abstracts, Siggraph 2001, Los Angeles, p 230

  26. Hämäläinen P, Ilmonen T, Höysniemi J, Lindholm M, Nykänen A (2005) Martial arts in artificial reality. In: Proceedings of CHI 2005 conference on human factors in computing systems, Portland, pp 781–790

  27. Henrysson A, Billinghurst M, Ollila M (2005) Face to face collaborative AR on mobile phones. In: Proceedings of the international symposium on mixed and augmented reality (ISMAR 2005), Vienna

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Eva Eriksson.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Eriksson, E., Hansen, T.R. & Lykke-Olesen, A. Movement-based interaction in camera spaces: a conceptual framework. Pers Ubiquit Comput 11, 621–632 (2007). https://doi.org/10.1007/s00779-006-0134-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00779-006-0134-z

Keywords

Navigation