Skip to main content

A Systematic Procedure for Comparing Template-Based Gesture Recognizers

  • Conference paper
  • First Online:
HCI International 2022 - Late Breaking Papers. Multimodality in Advanced Interaction Environments (HCII 2022)

Abstract

To consistently compare gesture recognizers under identical conditions, a systematic procedure for comparative testing should investigate how the number of templates, the number of sampling points, the number of fingers, and their configuration with other hand parameters such as hand joints, palm, and fingertips impact performance. This paper defines a systematic procedure for comparing recognizers using a series of test definitions, i.e. an ordered list of test cases with controlled variables common to all test cases. For each test case, its accuracy is measured by the recognition rate and its responsiveness by the execution time. This procedure is applied on six state-of-the-art template-based gesture recognizers on SHREC2019, a gesture dataset that contains simple and complex hand gestures tested and is largely used in the literature for competition in a user-independent scenario, and on Jackknife-lm, another challenging dataset. The results of the procedure identify the configurations in which each recognizer is the most accurate or the fastest.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Abraham, L., Urru, A., Norman, N., Wilk, M.P., Walsh, M.J., O’Flynn, B.: Hand tracking and gesture recognition using lensless smart sensors. Sensors 18(9), 2834 (2018). https://doi.org/10.3390/s18092834

    Article  Google Scholar 

  2. Akl, A., Valaee, S.: Accelerometer-based gesture recognition via dynamic-time warping, affinity propagation, & compressive sensing. In: 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 2270–2273 (2010). https://doi.org/10.1109/ICASSP.2010.5495895

  3. Aliofkhazraei, M., Ali, N.: Recent developments in miniaturization of sensor technologies and their applications. In: Comprehensive Materials Processing, pp. 245–306. Elsevier, Oxford (2014). https://doi.org/10.1016/B978-0-08-096532-1.01309-1

  4. Anthony, L., Wobbrock, J.O.: A lightweight multistroke recognizer for user interface prototypes. In: Proceedings of Graphics Interface 2010 (GI 2010), pp. 245–252. Canadian Information Processing Society, Toronto (2010). https://dl.acm.org/doi/10.5555/1839214.1839258

  5. Anthony, L., Wobbrock, J.O.: \$N-ProTractor: a fast and accurate multistroke recognizer. In: Proceedings of Graphics Interface 2012 (GI 2012), pp. 117–120. Canadian Information Processing Society, Toronto (2012). https://dl.acm.org/doi/10.5555/2305276.2305296

  6. Aquino, N., Vanderdonckt, J., Pastor, O.: Transformation templates: adding flexibility to model-driven engineering of user interfaces. In: Shin, S.Y., Ossowski, S., Schumacher, M., Palakal, M.J., Hung, C. (eds.) Proceedings of the 2010 ACM Symposium on Applied Computing (SAC), Sierre, Switzerland, 22–26 March 2010, pp. 1195–1202. ACM (2010). https://doi.org/10.1145/1774088.1774340

  7. Benitez-Garcia, G., Haris, M., Tsuda, Y., Ukita, N.: Finger gesture spotting from long sequences based on multi-stream recurrent neural networks. Sensors 20(2), 528 (2020). https://doi.org/10.3390/s20020528

    Article  Google Scholar 

  8. Brunelli, R.: Template Matching Techniques in Computer Vision: Theory and Practice. Wiley, New York (2009)

    Book  Google Scholar 

  9. Caputo, F.M., et al.: A 3 cent recognizer: simple and effective retrieval and classification of mid-air gestures from single 3D traces. In: Proceedings of the Conference on Smart Tools and Applications in Computer Graphics (STAG 2017), p. 9–15. Eurographics Association, Goslar, DEU (2017). https://doi.org/10.2312/stag.201

  10. Caputo, F.M., et al.: Online gesture recognition. In: Biasotti, S., Lavoué, G., Veltkamp, R. (eds.) Eurographics Workshop on 3D Object Retrieval, pp. 93–102. The Eurographics Association (2019). https://doi.org/10.2312/3dor.20191067

  11. Coyette, A., Schimke, S., Vanderdonckt, J., Vielhauer, C.: Trainable sketch recognizer for graphical user interface design. In: Baranauskas, C., Palanque, P., Abascal, J., Barbosa, S.D.J. (eds.) INTERACT 2007. LNCS, vol. 4662, pp. 124–135. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-74796-3_14

  12. Davis, A.: Getting Started with the Leap Motion SDK (2014). https://blog.leapmotion.com/getting-started-leap-motion-sdk/

  13. Ferrer, G., Sanfeliu, A.: Comparative analysis of human motion trajectory prediction using minimum variance curvature. In: Proceedings of the 6th International Conference on Human-Robot Interaction (HRI 2011), pp. 135–136. Association for Computing Machinery, New York (2011). https://doi.org/10.1145/1957656.1957698

  14. Filho, I.A.S., Chen, E.N., da Silva Junior, J.M., da Silva Barboza, R.: Gesture recognition using leap motion: a comparison between machine learning algorithms. In: ACM SIGGRAPH 2018 Posters (SIGGRAPH 2018). Association for Computing Machinery, New York (2018). https://doi.org/10.1145/3230744.3230750

  15. Khalaf, A.S., Alharthi, S.A., Dolgov, I., Toups, Z.O.: A comparative study of hand gesture recognition devices in the context of game design. In: Proceedings of the 2019 ACM International Conference on Interactive Surfaces and Spaces (ISS 2019), pp. 397–402. Association for Computing Machinery, New York (2019). https://doi.org/10.1145/3343055.3360758

  16. Khan, R.Z.: Comparative study of hand gesture recognition system. Comput. Sci. Inf. Technol. 2, 203–213 (2012). https://doi.org/10.5121/csit.2012.2320

    Article  Google Scholar 

  17. Leach, R.J.: Introduction to Software Engineering, 2nd edn. Chapman & Hall/CRC (2016)

    Google Scholar 

  18. Magrofuoco, N., Pérez-Medina, J.L., Roselli, P., Vanderdonckt, J., Villarreal, S.: Eliciting contact-based and contactless gestures with radar-based sensors. IEEE Access 7, 176982–176997 (2019). https://doi.org/10.1109/ACCESS.2019.2951349

    Article  Google Scholar 

  19. Magrofuoco, N., Roselli, P., Vanderdonckt, J.: Two-dimensional stroke gesture recognition: a survey. ACM Comput. Surv. 54(7), 155:1–155:36 (2022). https://doi.org/10.1145/3465400

  20. Marin, G., Dominio, F., Zanuttigh, P.: Hand gesture recognition with jointly calibrated Leap Motion and depth sensor. Multim. Tools Appl. 75(22), 14991–15015 (2015). https://doi.org/10.1007/s11042-015-2451-6

    Article  Google Scholar 

  21. Nielsen, J.: Usability Engineering, Interactive Technologies. Elsevier Science (1994)

    Google Scholar 

  22. Ousmer, M., Sluÿters, A., Magrofuoco, N., Roselli, P., Vanderdonckt, J.: Recognizing 3D trajectories as 2D multi-stroke gestures. In: Proceedings of the ACM on Human-Computer Interaction 4(ISS) (2020). https://doi.org/10.1145/3427326

  23. Rautaray, S.S., Agrawal, A.: Vision based hand gesture recognition for human computer interaction: a survey. Artif. Intell. Rev. 43(1), 1–54 (2012). https://doi.org/10.1007/s10462-012-9356-9

    Article  Google Scholar 

  24. Taranta, II, E.M., LaViola, Jr., J.J.: Penny pincher: a blazing fast, highly accurate \$-family recognizer. In: Proceedings of the 41st Graphics Interface Conference (GI 2015), pp. 195–202. Canadian Information Processing Society, Toronto (2015). https://dl.acm.org/doi/10.5555/2788890.2788925

  25. Taranta II, E.M., Samiei, A., Maghoumi, M., Khaloo, P., Pittman, C.R., LaViola Jr., J.J.: Jackknife: a reliable recognizer with few samples and many modalities. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI 2017 ), pp. 5850–5861. ACM, New York (2017). https://doi.org/10.1145/3025453.3026002

  26. Vanderdonckt, J., Roselli, P., Pérez-Medina, J.L.: !FTL, an articulation-invariant stroke gesture recognizer with controllable position, scale, and rotation invariances. In: Proceedings of the ICMI 2018, pp. 125–134. ACM, New York (2018). https://doi.org/10.1145/3242969.3243032

  27. Vatavu, R.D.: Improving gesture recognition accuracy on touch screens for users with low vision. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI 2017), pp. 4667–4679. ACM, New York (2017). https://doi.org/10.1145/3025453.3025941

  28. Vatavu, R.D., Anthony, L., Wobbrock, J.O.: Gestures as point clouds: a \$P recognizer for user interface prototypes. In: Proceedings of the 14th ACM International Conference on Multimodal Interaction (ICMI 2012), pp. 273–280. ACM, New York (2012). https://doi.org/10.1145/2388676.2388732

  29. Vatavu, R.D., Anthony, L., Wobbrock, J.O.: \$Q: a super-quick, articulation-invariant stroke-gesture recognizer for low-resource devices. In: Proceedings of the MobileHCI 2018, pp. 23:1–23:12. ACM, New York (2018). https://doi.org/10.1145/3229434.3229465

  30. Wang, A., Chen, G., Yang, J., Zhao, S., Chang, C.Y.: A comparative study on human activity recognition using inertial sensors in a smartphone. IEEE Sens. J. 16, 1 (2016). https://doi.org/10.1109/JSEN.2016.2545708

    Article  Google Scholar 

  31. Wang, C., Liu, Z., Chan, S.C.: Superpixel-Based Hand Gesture Recognition With Kinect Depth Camera. IEEE Transactions on Multimedia 17(1), 29–39 (2015). https://doi.org/10.1109/TMM.2014.2374357

    Article  Google Scholar 

  32. Wobbrock, J.O., Wilson, A.D., Li, Y.: Gestures without libraries, toolkits or training: a \$1 recognizer for user interface prototypes. In: Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology (UIST 2007), pp. 159–168. ACM, New York (2007). https://doi.org/10.1145/1294211.1294238

  33. Yasen, M., Jusoh, S.: A systematic review on hand gesture recognition techniques, challenges and applications. PeerJ Comput. Sci. 5, e218 (2019). https://doi.org/10.7717/peerj-cs.218

    Article  Google Scholar 

  34. Zengeler, N., Kopinski, T., Handmann, U.: Hand gesture recognition in automotive human-machine interaction using depth cameras. Sensors 19(1), 59 (2019). https://doi.org/10.3390/s19010059

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mehdi Ousmer .

Editor information

Editors and Affiliations

Appendices

A The New Recognizers Considered in the Experiment

1.1 A.1 $P3+X Recognizer

A variant of \(\textsf {\$}\textrm{P}^{3}+\) [27], that takes into account the direction-invariance by tracking conflicting templates (i.e., templates of the same gesture but performed in different directions). If a gesture matches with a conflicting template, its direction is compared with the direction of each conflicting template and the nearest one is chosen.

1.2 B.2 PennyPincher3D Recognizer

PennyPincher3D is an adaptation of the 2D recognizer PennyPincher [24]. The gestures are represented as a set of \(N-1\) vectors linking between N equidistant points. The recognizer matches the candidate gesture with the template that maximizes a dissimilarity score, computed as the sum of the angles between the vectors. The computation relies on basic mathematical operations as additions and multiplications. The gestures require just a resampling as prepossessing. This recognizer is scale- and position- invariant as most of the $-recognizers.

B The Datasets Considered in the Experiment

1.1 C.3 SHREC2019 [10]

The SHREC2019 dataset [10] contains a sequence of 3D points and quaternions for each hand’s joint designating one of five gesture classes (Fig. 8): “Cross” (X), “Circle” (O), “V-mark” (V), “Caret” (/\(\backslash \)), and “Square” ([])). It served in (SHREC) track, a contest on online gesture recognition to detect command gestures from hands’ movements in a virtual reality context. The proposed dataset consists of 195 3D movements performed by 13 participants with the whole hand. The dataset contains unsegmented gestures, the training set and the testing set were merged to create a unique dataset in which, unnecessary hand movements were removed from the gestures.

Fig. 8.
figure 8

The SHREC2019 gesture classes.

1.2 D.4 Jackknife-LM [25]

The Jackknife-LM (Jackknife-LeapMotion) dataset [25] contains 3D complex gestures of the hand and saved as 3D skeleton which is provided by the LeapMotion. We used the segmented gestures composed by 360 samples of 9 different gesture classes for example “Fist Circles”, “Snip Snip”, “Explode” (Fig. 9). It was used to test a rejection approach of non-gesture sequences from a continuous data stream. While segmented gestures make up the training set, authors employ unsegmented sessions of samples in the test [25].

Fig. 9.
figure 9

The Jackknife-LM gesture classes [25].

C Recognition Rates

1.1 E.5 Recognition Rate for Best Condition for the SHREC2019

Fig. 10.
figure 10

Recognition rate (left) and the number of recognized gestures per class over 100 trials (right) of all recognizers for user independent scenario, for the optimal conditions defined by the de Borda ranking: \(A = 2\,(T + I)\), \(N = 32\), \(T = 16\). Error bars show a confidence interval of \(95\%\).

1.2 F.6 Recognition Rates Tables for the SHREC2019

figure a

1.3 G.7 Recognition Rates Tables for the Jackknife-LM

figure b

1.4 H.8 The Ranking of the Recognizers Based on the Best Individual Rates by Articulation for SHREC2019

Fig. 11.
figure 11

Ranking of the best individual recognition rates above 80% by number of joints (A) and number of points (N) for the SHREC2019 dataset.

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ousmer, M., Sluÿters, A., Magrofuoco, N., Roselli, P., Vanderdonckt, J. (2022). A Systematic Procedure for Comparing Template-Based Gesture Recognizers. In: Kurosu, M., et al. HCI International 2022 - Late Breaking Papers. Multimodality in Advanced Interaction Environments. HCII 2022. Lecture Notes in Computer Science, vol 13519. Springer, Cham. https://doi.org/10.1007/978-3-031-17618-0_13

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-17618-0_13

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-17617-3

  • Online ISBN: 978-3-031-17618-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics