ABSTRACT
This course will introduce participants to practical aspects of human gesture production, gesture recognition, and gesture analysis. Notions such as gesture training sets, user-dependent and user-independent training, and variability of gesture articulation will be clarified for the audience in the context of designing gesture user interfaces. Simple and efficient gesture recognition and gesture analysis techniques will be exemplified and accompanied by real-world data analysis with software tools readily available to download and use. An overview of gesture taxonomies and design guidelines will accompany the discussion. Participants will be able to try out gesture analysis on their own computers. It is my hope that this course will effect into a valuable source of information for researchers and practitioners entering the gesture field who are interested in prototyping novel interactive systems controllable by means of human gestures.
- Lisa Anthony, Radu-Daniel Vatavu, and Jacob O. Wobbrock. 2013. Understanding the consistency of users' pen and finger stroke gesture articulation. In Proc. of Graphics Interface 2013 (GI '13), 87--94.Google ScholarDigital Library
- Lisa Anthony and Jacob O. Wobbrock. 2010. A lightweight multistroke recognizer for user interface prototypes. In Proceedings of Graphics Interface 2010 (GI '10). Canadian Information Processing Society, Toronto, Ont., Canada, Canada, 245--252.Google ScholarDigital Library
- Caroline Appert and Shumin Zhai. 2009. Using strokes as command shortcuts: cognitive benefits and toolkit support. In Proc. of the SIGCHI Conf. on Human Factors in Computing Systems (CHI '09). ACM, New York, NY, USA, 2289--2298. Google ScholarDigital Library
- Xiang Cao and Shumin Zhai. 2007. Modeling human performance of pen stroke gestures. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '07). ACM, New York, NY, USA, 1495--1504. Google ScholarDigital Library
- Jean Ann Graham, Michael Argyle. 1975. A crosscultural study of the communication of extra-verbal meaning by gestures. International Journal of Psychology 10, 57--67. Google ScholarCross Ref
- Adam Kendon. 1994. Do gestures communicate? A review. Research on Language and Social Interaction 27, 175--200. Google ScholarCross Ref
- Jana M. Iverson, Susan Goldin-Meadow. 1998. Why people gesture when they speak. Nature, 396:228. Google ScholarCross Ref
- David McNeill. 1992. Hand and Mind: What Gesture Reveals about Thought. University Chicago PressGoogle Scholar
- Miguel A. Nacenta, Yemliha Kamber, Yizhou Qiang, and Per Ola Kristensson. 2013. Memorability of pre-designed and user-defined gesture sets. In Proc. of the SIGCHI Conference on Human Factors in Computing Systems (CHI '13). ACM, 1099--1108. Google ScholarDigital Library
- M. Nielsen, M. Störring, T.B. Moeslund, E. Granum. 2004. A procedure for developing intuitive and ergonomic gesture interfaces for HCI. Int'l Gesture Workshop 2003, LNCS vol. 2915. Heidelberg: Springer-Verlag, 409--420. Google ScholarCross Ref
- Yosra Rekik, Radu-Daniel Vatavu, and Laurent Grisoni. 2014. Understanding Users' Perceived Difficulty of Multi-Touch Gesture Articulation. In Proc. of the 16th Int. Conf. on Multimodal Interaction (ICMI '14). ACM, 232--239. Google ScholarDigital Library
- Yosra Rekik, Radu-Daniel Vatavu, and Laurent Grisoni. 2014. Match-up & conquer: a two-step technique for recognizing unconstrained bimanual and multi-finger touch input. In Proc. of the 2014 Int. Working Conf. on Advanced Visual Interfaces (AVI '14). ACM, New York, NY, USA, 201--208. Google ScholarDigital Library
- Radu-Daniel Vatavu and Jacob O. Wobbrock. 2015. Formalizing Agreement Analysis for Elicitation Studies: New Measures, Significance Test, and Toolkit. In Proc. of the 33rd Annual ACM Conf. on Human Factors in Computing Systems (CHI '15). ACM, New York, NY, USA, 1325--1334. Google ScholarDigital Library
- Radu-Daniel Vatavu and Jacob O. Wobbrock. 2016. Between-Subjects Elicitation Studies: Formalization and Tool Support. In Proc. of the 2016 Conf. on Human Factors in Computing Systems (CHI '16). ACM, USA, 3390--3402. Google ScholarDigital Library
- Radu-Daniel Vatavu, Lisa Anthony, and Jacob O. Wobbrock. 2014. Gesture Heatmaps: Understanding Gesture Performance with Colorful Visualizations. In Proc. of the 16th Int. Conf. on Multimodal Interaction (ICMI '14). ACM, 172--179. Google ScholarDigital Library
- Radu-Daniel Vatavu, Lisa Anthony, and Jacob O. Wobbrock. 2013. Relative accuracy measures for stroke gestures. In Proc. of the 15th ACM on International conference on multimodal interaction (ICMI '13). ACM, New York, NY, USA, 279--286. Google ScholarDigital Library
- Radu-Daniel Vatavu, Lisa Anthony, and Jacob O. Wobbrock. 2012. Gestures as point clouds: a $P recognizer for user interface prototypes. In Proc. of the 14th ACM Int. Conf. on Multimodal interaction (ICMI '12). ACM, USA, 273--280. Google ScholarDigital Library
- Radu-Daniel Vatavu, Gabriel Cramariuc, Doina Maria Schipor. 2015. Touch Interaction for Children Aged 3 to 6 Years: Experimental Findings and Relationship to Motor Skills. International Journal of Human-Computer Studies 74, 54--76. Google ScholarDigital Library
- Radu-Daniel Vatavu, Daniel Vogel, Géry Casiez, and Laurent Grisoni. 2011. Estimating the perceived difficulty of pen gestures. In Proc. of the 13th IFIP TC 13 Int. Conf. on Human-computer interaction (INTERACT'11), 89--106. Google ScholarCross Ref
- Radu-Daniel Vatavu. 2013. The Impact of Motion Dimensionality and Bit Cardinality on the Design of 3D Gesture Recognizers. International Journal of Human-Computer Studies, 71(4), 387--409. Google ScholarDigital Library
- Radu-Daniel Vatavu. 2017. Improving Gesture Recognition Accuracy on Touch Screens for Users with Low Vision. In Proc. of CHI '17, the 35th ACM Conf. on Human Factors in Computing SystemsGoogle ScholarDigital Library
- Radu-Daniel Vatavu, Annette Mossel, and Christian Schönauer. 2016. Digital vibrons: understanding users' perceptions of interacting with invisible, zero-weight matter. In Proc. of the 18th Int. conf. on Human-Computer Interaction with Mobile Devices and Services (MobileHCI '16), 217--226 Google ScholarDigital Library
- Jacob O. Wobbrock, Andrew D. Wilson, and Yang Li. 2007. Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes. In Proc. of the 20th annual ACM Symp. on User interface software and technology (UIST '07). ACM, New York, NY, USA, 159--168.Google ScholarDigital Library
- Jacob O. Wobbrock, Htet Htet Aung, Brandon Rothrock, and Brad A. Myers. 2005. Maximizing the guessability of symbolic input. In CHI '05 Extended Abstracts on Human Factors in Computing Systems (CHI EA '05), 1869--1872. Google ScholarDigital Library
- Jacob O. Wobbrock, Meredith Ringel Morris, and Andrew D. Wilson. 2009. User-defined gestures for surface computing. In Proc. of the SIGCHI Conf. on Human Factors in Computing Systems (CHI '09). ACM, New York, NY, USA, 1083--1092. Google ScholarDigital Library
Index Terms
- Fundamentals of Gesture Production, Recognition, and Analysis
Recommendations
Designing, Engineering, and Evaluating Gesture User Interfaces
CHI EA '18: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing SystemsThis course will introduce participants to the three main stages of the development life cycle of gesture-based interactions: (ul) how to design a gesture user interface (UI) by carefully considering key aspects, such as gesture recognition techniques, ...
Multi-scenario gesture recognition using Kinect
CGAMES '12: Proceedings of the 2012 17th International Conference on Computer Games: AI, Animation, Mobile, Interactive Multimedia, Educational & Serious Games (CGAMES)Hand gesture recognition (HGR) is an important research topic because some situations require silent communication with sign languages. Computational HGR systems assist silent communication, and help people learn a sign language. In this article, a ...
iFAD Gestures: Understanding Users’ Gesture Input Performance with Index-Finger Augmentation Devices
CHI '23: Proceedings of the 2023 CHI Conference on Human Factors in Computing SystemsWe examine gestures performed with a class of input devices with distinctive quality properties in the wearables landscape, which we call “index-Finger Augmentation Devices” (iFADs). We introduce a four-level taxonomy to characterize the diversity of ...
Comments