skip to main content
10.1145/3340555.3356092acmotherconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
abstract

Tailoring Motion Recognition Systems to Children’s Motions

Published:14 October 2019Publication History

ABSTRACT

Motion-based applications are becoming increasingly popular among children and require accurate motion recognition to ensure meaningful interactive experiences. However, motion recognizers are usually trained on adults’ motions. Children and adults differ in terms of their body proportions and development of their neuromuscular systems, so children and adults will likely perform motions differently. Hence, motion recognizers tailored to adults will likely perform poorly for children. My PhD thesis will focus on identifying features that characterize children’s and adults’ motions. This set of features will provide a model that can be used to understand children’s natural motion qualities and will serve as the first step in tailoring recognizers to children’s motions. This paper describes my past and ongoing work toward this end and outlines the next steps in my PhD work.

References

  1. Aishat Aloba, Gianne Flores, Julia Woodward, Alex Shaw, Amanda Castonguay, Isabella Cuba, Yuzhu Dong, Eakta Jain, and Lisa Anthony. 2018. Kinder-Gator: The UF kinect database of child and adult motion.. In EG 2018 - Short Papers., Olga Diamanti and Amir Vaxman (Eds.). 13–16. https://doi.org/10.2312/egs.20181033Google ScholarGoogle Scholar
  2. Aishat Aloba, Annie Luc, Yuzhu Dong, Rong Zhang, Eakta Jain, and Lisa Anthony. 2019. Quantifying differences between child and adult motion based on gait features. In International Conference on Human-Computer Interaction (HCII ’19). 385–402. https://doi.org/10.1007/978-3-030-23563-5_31Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Lisa Anthony, Quincy Brown, Jaye Nias, Berthel Tate, and Shreya Mohan. 2012. Interaction and recognition challenges in interpreting children’s touch and gesture input on mobile devices. In Proceedings of the 2012 ACM international conference on Interactive tabletops and surfaces (ITS ’12). ACM Press, 225. https://doi.org/10.1145/2396636.2396671Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Lisa Anthony and Jacob O Wobbrock. 2012. $ N-protractor: a fast and accurate multistroke recognizer. In Proceedings of Graphics Interface (GI ’12). Canadian Information Processing Society, 117–120.Google ScholarGoogle Scholar
  5. Donald J Berndt and James Clifford. 1994. Using dynamic time warping to find patterns in time series. In KDD Workshop on Knowledge Discovery in Databases. 359–370.Google ScholarGoogle Scholar
  6. Xiang Cao and Shumin Zhai. 2007. Modeling human performance of pen stroke gestures. In Proceedings of the SIGCHI conference on Human factors in computing systems (CHI ’07). ACM, 1495–1504.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Corinna Cortes and Vladimir Vapnik. 1995. Support-vector networks. Machine learning 20, 3 (1995), 273–297.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Thomas Cover and Peter Hart. 1967. Nearest neighbor pattern classification. IEEE transactions on information theory 13, 1 (1967), 21–27.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Janto F. Dreijer and Ben M. Herbst. 2008. Action classification using the average of pose changes. In Symposium of the Pattern Recognition Association of South Africa. 85–85.Google ScholarGoogle Scholar
  10. Tomasz Hachaj and Marek R. Ogiela. 2014. Rule-based approach to recognizing human body poses and gestures in real time. Multimedia Systems (2014), 81–99. https://doi.org/10.1007/s00530-013-0332-2Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. John A Hartigan and Manchek A Wong. 1979. Algorithm AS 136: A k-means clustering algorithm. Journal of the Royal Statistical Society. Series C (Applied Statistics) 28, 1(1979), 100–108.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Donald F. Huelke. 1998. An overview of anatomical considerations of infants and children in the adult world of automobile safety design. Association for the Advancement of Automotive Medicine 42 (1998), 93–113. https://doi.org/10.1145/982452.982461 arxiv:arXiv:1011.1669v3Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Wan Noorshahida Mohd Isa. 2005. Analysis on spatial and temporal features of gait kinematics. In Fourth IEEE Workshop on Automatic Identification Advanced Technologies (AutoID’05). IEEE, 130–133.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Eakta Jain, Lisa Anthony, Aishat Aloba, Amanda Castonguay, Isabella Cuba, Alex Shaw, and Julia Woodward. 2016. Is the motion of a child perceivably different from the motion of an adult?ACM Transactions on Applied Perception 13, 4 (jul 2016), 1–17. https://doi.org/10.1145/2947616Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Charlene Jennett, Anna L Cox, Paul Cairns, Samira Dhoparee, Andrew Epps, Tim Tijs, and Alison Walton. 2008. Measuring and defining the experience of immersion in games. International journal of human-computer studies 66, 9 (2008), 641–661.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Roanna Lun and Wenbing Zhao. 2015. A survey of applications and human motion recognition with microsoft kinect. International Journal of Pattern Recognition and Artificial Intelligence (2015). https://doi.org/10.1142/s0218001415550083Google ScholarGoogle ScholarCross RefCross Ref
  17. Microsoft. [n.d.]. Kinect for Windows. https://developer.microsoft.com/en-us/windows/kinectGoogle ScholarGoogle Scholar
  18. Jasmir Nijhar, Nadia Bianchi-Berthouze, and Gemma Boguslawski. 2012. Does movement recognition precision affect the player experience in exertion games?. In International Conference on Intelligent Technologies for Interactive Entertainment (INTETAIN ’12). 73–82. https://doi.org/10.1007/978-3-642-30214-5_9Google ScholarGoogle ScholarCross RefCross Ref
  19. Santiago Riofrío, David Pozo, Jorge Rosero, and Juan Vásques. 2017. Gesture recognition using dynamic time warping and kinect: A practical approach. In International Conference on Information Systems and Computer Science. 302–308.Google ScholarGoogle ScholarCross RefCross Ref
  20. Mikel D. Rodriguez, Javed Ahmed, and Mubarak Shah. 2008. Action MACH: A spatio-temporal maximum average correlation height filter for action recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR ’08). 6–6. https://doi.org/10.1109/CVPR.2008.4587727Google ScholarGoogle ScholarCross RefCross Ref
  21. S Rasoul Safavian and David Landgrebe. 1991. A survey of decision tree classifier methodology. IEEE transactions on systems, man, and cybernetics 21, 3(1991), 660–674.Google ScholarGoogle ScholarCross RefCross Ref
  22. Christian Schuldt, Ivan Laptev, and Barbara Caputo. 2004. Recognizing human actions: a local SVM approach. In Proceedings of the International Conference on Pattern Recognition (ICPR ’04), Vol. 3. IEEE, 32–36.Google ScholarGoogle ScholarCross RefCross Ref
  23. Alex Shaw and Lisa Anthony. 2016. Analyzing the articulation features of children’s touchscreen gestures. In ACM International Conference on Multimodal Interaction (ICMI ’16). 333–340. https://doi.org/10.1145/2993148.2993179Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. David H. Sutherland. 1997. The development of mature gait. Gait & Posture 6, 2 (1997), 163–170. https://doi.org/10.1016/S0966-6362(97)00029-5Google ScholarGoogle ScholarCross RefCross Ref
  25. Radu-Daniel Vatavu. 2017. Beyond features for recognition: human-readable measures to understand users’ whole-body gesture performance. International Journal of Human–Computer Interaction 33, 9(2017), 713–730.Google ScholarGoogle Scholar
  26. Jacob O. Wobbrock, Andrew D. Wilson, and Yang Li. 2007. Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes. In Proceedings of the ACM symposium on User interface software and technology (UIST ’07). https://doi.org/10.1145/1294211.1294238Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Julia Woodward, Alex Shaw, Annie Luc, Brittany Craig, Juthika Das, Phillip Hall Jr., Akshay Holla, Germaine Irwin, Danielle Sikich, Quincy Brown, and Lisa Anthony. 2016. Characterizing how interface complexity affects children’s touchscreen interactions. In Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems (CHI ’16). ACM Press, 1921–1933.Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Hee Deok Yang, A. Yeon Park, and Seong Whan Lee. 2006. Human-robot interaction by whole body gesture spotting and recognition. In International Conference on Pattern Recognition, Vol. 4. 774–777. https://doi.org/10.1109/ICPR.2006.642Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    ICMI '19: 2019 International Conference on Multimodal Interaction
    October 2019
    601 pages
    ISBN:9781450368605
    DOI:10.1145/3340555

    Copyright © 2019 Owner/Author

    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 14 October 2019

    Check for updates

    Qualifiers

    • abstract
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate453of1,080submissions,42%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format