skip to main content
10.1145/3406324.3410709acmotherconferencesArticle/Chapter ViewAbstractPublication PagesmobilehciConference Proceedingsconference-collections
extended-abstract

Quantifying the Consistency of Gesture Articulation for Users with Low Vision with the Dissimilarity-Consensus Method

Published:25 February 2021Publication History

ABSTRACT

We apply the dissimilarity-consensus method to quantify and report the articulation consistency of gestures produced on touchscreens by users with low vision, which we compare to the consistency of people without visual impairments. We report results in terms of dissimilarity-consensus growth curves and logistic models on a public dataset of 6,562 stroke-gestures collected from 54 participants, of which 27 with low vision. Our empirical results show that participants with low vision were 28% less consistent in their gesture articulations compared to the participants without visual impairments. We also demonstrate the suitability of the method, applied so far for whole-body gestures only, for the analysis of touchscreen stroke-gestures.

References

  1. Lisa Anthony, Radu-Daniel Vatavu, and Jacob O. Wobbrock. 2013. Understanding the Consistency of Users’ Pen and Finger Stroke Gesture Articulation. In Proceedings of Graphics Interface 2013 (Regina, Sascatchewan, Canada) (GI ’13). Canadian Information Processing Society, 87–94. https://dl.acm.org/doi/10.5555/2532129.2532145Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Bogdan-Florin Gheran, Jean Vanderdonckt, and Radu-Daniel Vatavu. 2018. Gestures for Smart Rings: Empirical Results, Insights, and Design Implications. In Proceedings of the 2018 Designing Interactive Systems Conference (Hong Kong, China) (DIS ’18). ACM, New York, NY, USA, 623–635. https://doi.org/10.1145/3196709.3196741Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Kasper Hornbæk, Søren S. Sander, Javier Andrés Bargas-Avila, and Jakob Grue Simonsen. 2014. Is Once Enough? On the Extent and Content of Replications in Human-Computer Interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Toronto, Ontario, Canada) (CHI ’14). ACM, New York, NY, USA, 3523–3532. https://doi.org/10.1145/2556288.2557004Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Shaun K. Kane, Jacob O. Wobbrock, and Richard E. Ladner. 2011. Usable Gestures for Blind People: Understanding Preference and Performance. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Vancouver, BC, Canada) (CHI ’11). ACM, New York, NY, USA, 413–422. https://doi.org/10.1145/1978942.1979001Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Luis A. Leiva, Daniel Martín-Albo, and Radu-Daniel Vatavu. 2017. Synthesizing Stroke Gestures Across User Populations: A Case for Users with Visual Impairments. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). ACM, New York, NY, USA, 4182–4193. https://doi.org/10.1145/3025453.3025906Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Maria-Doina Schipor and Radu-Daniel Vatavu. 2017. Coping Strategies of People with Low Vision for Touch Input: A Lead-in Study. In Proceedings of the 6th IEEE International Conference on e-Health and Bioengineering(EHB ’17). 357–360. http://dx.doi.org/10.1109/EHB.2017.7995435Google ScholarGoogle ScholarCross RefCross Ref
  7. Radu-Daniel Vatavu. 2017. Improving Gesture Recognition Accuracy on Touch Screens for Users with Low Vision. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). ACM, New York, NY, USA, 4667–4679. https://doi.org/10.1145/3025453.3025941Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Radu-Daniel Vatavu. 2019. The Dissimilarity-Consensus Approach to Agreement Analysis in Gesture Elicitation Studies. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). ACM, New York, NY, USA, 1–13. https://doi.org/10.1145/3290605.3300454Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Radu-Daniel Vatavu, Lisa Anthony, and Jacob O. Wobbrock. 2012. Gestures as Point Clouds: A $P Recognizer for User Interface Prototypes. In Proceedings of the 14th ACM International Conference on Multimodal Interaction (Santa Monica, California, USA) (ICMI ’12). ACM, New York, NY, USA, 273–280. https://doi.org/10.1145/2388676.2388732Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Radu-Daniel Vatavu, Bogdan-Florin Gheran, and Maria-Doina Schipor. 2018. The Impact of Low Vision on Touch-Gesture Articulation on Mobile Devices. IEEE Perv. Comp. 17, 1 (2018), 27–37. https://doi.org/10.1109/MPRV.2018.011591059Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Radu-Daniel Vatavu and Ovidiu-Ciprian Ungurean. 2019. Stroke-Gesture Input for People with Motor Impairments: Empirical Results & Research Roadmap. In Proc. of the 2019 CHI Conf. on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). ACM, New York, NY, USA, 1–14. https://doi.org/10.1145/3290605.3300445Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Radu-Daniel Vatavu and Jean Vanderdonckt. 2020. What Gestures Do Users with Visual Impairments Prefer to Interact with Smart Devices? And How Much We Know About It. In Companion Publication of the 2020 ACM Designing Interactive Systems Conference (Eindhoven, Netherlands) (DIS’ 20 Companion). ACM, New York, NY, USA, 85–90. https://doi.org/10.1145/3393914.3395896Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Jacob O. Wobbrock, Meredith Ringel Morris, and Andrew D. Wilson. 2009. User-Defined Gestures for Surface Computing. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Boston, MA, USA) (CHI ’09). ACM, New York, NY, USA, 1083–1092. https://doi.org/10.1145/1518701.1518866Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Jacob O. Wobbrock, Andrew D. Wilson, and Yang Li. 2007. Gestures without Libraries, Toolkits or Training: A $1 Recognizer for User Interface Prototypes. In Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology (Newport, Rhode Island, USA) (UIST ’07). ACM, New York, NY, USA, 159–168. https://doi.org/10.1145/1294211.1294238Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    MobileHCI '20: 22nd International Conference on Human-Computer Interaction with Mobile Devices and Services
    October 2020
    248 pages
    ISBN:9781450380522
    DOI:10.1145/3406324

    Copyright © 2020 Owner/Author

    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 25 February 2021

    Check for updates

    Qualifiers

    • extended-abstract
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate202of906submissions,22%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format