Abstract
This paper describes the development of a new technique for touchscreen interaction, based on a single gesture-driven adaptive software button. The button is intended to substitute the software keyboard, and provides text-entry functionality. Input is accomplished through recognition of finger gestures that is comprised of movement towards the eight basic directions in any position. The target user group of such an interaction technique is primarily blind people who could benefit significantly. The adaptability of the button provides complementary help and follows the style of interaction in a natural way. The analysis of the results, collected from twelve blindfolded subjects, revealed an encouraging tendency. During blind manipulation on touch screen, three of the subjects achieved a maximal typing speed of about 12 wpm after five trials. This suggests that the technique developed is reliable and robust enough to be possibly applied to diverse application platforms, including personal device assistants.
Similar content being viewed by others
References
AT&T Natural Voices Text-to-Speech Engine. Product information is available at: http://www.naturalvoices.att.com
British Blind Sport (2004) Web site http://www.britishblindsport.org.uk/archery.htm
Bourhis G, Pino P (2003) Temporal analysis of the use of an augmentative communication device. In: Proceedings of AAATE’03. IOS Press, Netherlands, pp 294–298
Brewster SA, Brown LM (2004) Tactons: structured tactile messages for non-visual information display. In: Proceedings of Australasian user interface conference 2004. Australian Computer Society, Dunedin, pp 15–23
Callahan J, Hopkins D, Weiser M, Shneiderman B (1988) A comparative analysis of pie menu performance. In: Proceedings of CHI’88 conference, Washington DC. Web site (2004) http://www.catalog.com/hopkins/piemenus/callahan.ps
Card SK, Mackinlay JD, Robertson GG (1991) A morphological analysis of the design space of input devices. ACM Trans Inf Syst 9:99–122
Evreinov G, Raisamo R (2002) Tactile pointer for touchscreen manipulations. In: Proceedings of HANDICAP 2002. L’Institut Federativf de Recherche sur les Aides Techniques pour Personnes Handicapees, Paris, pp 115–120
Geißler J (1995) Gedrics: the next generation of icons. In: Proceedings of the INTERACT’95, Lillehammer Norway, pp 73–78
Gnatenko V (2004) Multi-directional Input Keypad (the MIK)—text input solution for mobile devices. Web site http://www.vitgn.com/
Hansen JP, Johansen AS, Hansen DW, Kenji I, Satoru M (2003) Command without a click: dwell time typing by mouse and gaze selections. In: Proceedings of INTERACT’03. IOS Press, IFIP, pp 121–128
Jhaveri N (2003) Two characters per stroke—a novel pen-based text input technique. In: Grigori Evreinov (ed) New interaction techniques 2003 (Report B-2003-5), Department of Computer and Information Sciences, University of Tampere, pp 10–15. Web site (2004) http://www.cs.uta.fi/reports/bsarja/B-2003-5.pdf
Isokoski P (2004) Performance of menu-augmented soft keyboards. In: CHI 2004, ACM conference on human factors in computing systems. CHI Letters 6(1):423–430
Kamel HM, Landay JA (2000) A study of blind drawing practice: creating graphical information without the visual channel. In: Proceedings of ASSETS 2000, pp 34–41
Lesher GW, Higginbotham DJ, Moulton BJ (2000) Techniques for automatically updating scanning delays. In: Proceedings of the RESNA 2000, pp 85–87
MacKenzie IS (2001) A note on phrase sets for evaluating text entry techniques. Web site (2004) http://www.yorku.ca/mack/RN-PhraseSet.html
Moyle M, Cockburn A (2002) Analysing mouse and pen flick gestures. In: Proceedings of the SIGCHI-NZ symposium on computer-human interaction, Hamilton, pp 19–24
Nesbat SB (2003) A system for fast, full text entry for small electronic devices. In: Proceedings of ICMI’03. ACM Press, pp 4–11
Partridge K, Chatterjee S, Sazawal V, Borriello G, Want R (2002) TiltType Accelerometer-supported text entry for very small devices. In: Proceedings of ACM UIST symposium on user interface software and technology. ACM Press, paris, pp 201–204
Royal Blind Society (2004) Web site http://www.rbs.org.au
Rönnberg J, Samuelsson E, Borg E (2002) Exploring the perceived world of the deaf-blind: on the development of an instrument. Int J Audiol 41(1):136–143
Simpson RC, Koester HH (1999) Adaptive one-switch row-column scanning. IEEE Trans Rehabil Eng 7(4):464–473
Stephanidis C, Savidis A (2001) Interface development toolkits for non-visual and switch-based interaction. In: Proceedings of ERCIM News, Special Theme. Human Computer Interaction 46:4–15
The SIMS Video Game (2004) Electronic Arts, Web site (2004) http://www.thesims.com
Vanderheiden GC (1996) Use of audio-haptic interface techniques to allow nonvisual access to touchscreen appliances. In: Proceedings of human factors and ergonomics society annual conference. Web site (2004) http://www.tracecenter.org/docs/touchscreen/chi_conf.htm
Venolia D, Neiburg F (1994) T-cube: a fast, self-disclosing pen-based alphabet. In: Proceedings of CHI’94, ACM Press, pp 265–270
Word Frequencies (2004) University of Tampere, Web site http://www.cs.uta.fi/%7Escott/mmm/WordFrequencies.xls
Acknowledgements
This work was financially supported by the Academy of Finland (grant 200761), and by the Nordic Development Centre for Rehabilitation Technology (NUH).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Yfantidis, G., Evreinov, G. Adaptive blind interaction technique for touchscreens. Univ Access Inf Soc 4, 328–337 (2006). https://doi.org/10.1007/s10209-004-0109-7
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10209-004-0109-7