Abstract
This paper discusses the design and initial evaluation of a visual on-screen keyboard, operated by eye-gaze, intended for use by motor-impaired users. The idea of an on-screen keyboard controlled by eye or by other modalities is not new. However, the keyboard presented here is different in two important respects. First, it enables interaction with unmodified standard Graphical User Interface (GUI) software written for able-bodied users, and provides eye-based control over menus, dialogue boxes, and scrollers; it is not solely designed around the need to enter text. Second, the software architecture enables the keyboard to respond to events generated in the windows environment by the application it is controlling. This allows the keyboard to adapt automatically to the application context by, for example, loading a specific set of keys designed for use with particular menus whenever a menu is displayed in the target application. Results of initial evaluation trials are presented and the implications for improvements in design are discussed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bishop, J. B. & Myers, G. A. (1993), “Development of an Effective Computer Interface for Persons with Mobility Impairment”, Proceedings of the Annual Conference on Engineering in Medicine and Biology 15 (3), 1266–1267.
Cleveland, D. & Cleveland, N. (1992), Eyegaze Eyetracking System, in “Imagina: Images Beyond Imagination, Proceedings of the 11th Monte-Carlo International Forum on New Images”, pp. 1–7.
Demasco, P. W. & McKoy, K. F. (1992), “Generating Text From Compressed Input: An Intelligent Interface for People with Severe Motor Impairments”, Communications of the ACM 35 (5), 68–77.
Douglas, S. D. & Happ, A. J. (1993), Evaluating Performance, Discomfort, and Subjective Preference between Computer Keyboard Designs, in G. Salvendy and M. J. Smith (eds.), “Proceedings of the 5th International Conference on Human–Computer Interaction (HCI International ‘83)”, Elsevier Science, pp. 1064–1072.
Eulenberg, J. B., King, M. T. & Patterson, H. (1985), Eye-Controlled Communication, in “Proceedings of Speech Technology ‘85 — Voice Input and Output Applications”, pp.210–211.
Frey, L. A., White, K. P. & Hutchinson, T. E. (1990), “Eye-Gaze Word Processing”, IEEE Transactions in Systems, Man and Cybernetics 20 (4), 944–950.
Friedman, M. B., Kiliany, G. & Dzmura, M. (1985), An Eye Gaze Controlled Keyboard, in “Proceedings of the 2nd International Conference on Rehabilitation Engineering”, pp.446–447.
Friedman, M. B., Kiliany, G., Dzmura, M. & Anderson, D. (1981), The Eyetracker Communication System, in “1st National Search for Applications to Aid the Handicapped”, John Hopkins University, pp. 183–185.
Gould, J. D., Lewis, C. & Barnes, V. (1985), Effects of Cursor Speed on Text-Editing, in “Proceedings of CHI’85: Human Factors in Computing Systems”, ACM Press, pp. 710.
Hutchinson, T. E., White, K. P., Martin, W. N., Reichert, K. C. & Frey, L. A. (1989), “Human–Computer Interaction Using Eye-Gaze Input”, IEEE Transactions in Systems, Man and Cybernetics 19 (6), 1527–1534.
Istance, H. O. & Howarth, P. A. (1994), Keeping an Eye on your Interface: The Potential for Eye-based Control of Graphical User Interfaces (GUI’s), in G. Cockton, S. Draper and G. Wier (eds.), “People and Computers IX (Proceedings of HCI’94)”, Cambridge University Press, pp. 195–209.
Jacob, R. J. K. (1993), Eye Movement-based Human–Computer Interaction Techniques: Toward Non-command Interfaces, in R. Hartson and D. Hix (eds.), “Advances in Human–Computer Interaction”, Vol. 4, Ablex, pp. 151–190.
Koester, H. H. & Levine, S. P. (1994), “Modeling the Speed of Text Entry with a Word Prediction Interface”, IEEE Transactions on Rehabilitation Engineering 2 (3), 177–187.
Mackenzie, S., Nonnecke, B., Riddersma, S., McQueen, C. & Meltz, M. (1994), “Alphanumeric Entry on Pen-based Computers”, International Journal of Human–Computer Studies 41 (5), 775–792.
Murphy, R. A. & Basili, A. (1993), Developing the User–System Interface for a Communications System for ALS Patients and Others with Severe Neurological Impairments, in “Designing for Diversity; Proceedings of the Human Factors and Ergonomics Society”, Vol. 2, Human Factors and Ergonomics Society, pp. 854–858.
Nantais, T., Shein, F. & Treviranus, J. (1994), “A Predictive Selection Technique for Single-digit Typing With a Visual Keyboard”, IEEE Transactions on Rehabilitation Engineering 2 (3), 130–134.
Quill, L. L. & Biers, D. W. (1993), On-screen Keyboards: Which Arrangements Should Be Used?, in “Designing for Diversity; Proceedings of the Human Factors and Ergonomics Society”, Vol. 2, Human Factors and Ergonomics Society, pp. 1142–1146.
Roussos, P. (1992), Effects of Keyboard Layout on Children’s Performance and Interaction with Computers, Master’s thesis, University of Leeds.
Schofield, J. M. (1981), Chapter ?, in P. A. Samet (ed.), “Microcomputer-based Aids for the Disabled”, Heyden.
Shaw, R., Crisman, E., Loomis, A. & Laszewski, Z. (1990), The Eye-Wink Control Interface: Using the Computer to Provide the Severely Disabled with Increased Flexibility and Comfort, in “Proceedings of the 3rd Annual IEEE Symposium on Computer-Based Medical Systems (CBMS’90)”, IEEE Computer Society Press, pp.105–1 l l.
Shein, F., Hamann, G., Brownlow, N., Treviranus, J., Milner, M. & Parnes, P. (1991), WIVIK: A Visual Keyboard for Windows 3.0, in “Proceedings of the 14th Annual Conference of the Rehabilitation Engineering Society of North America (RESNA)”, pp.160–162.
Shein, F., Treviranus, J., Hamann, G., Galvin, R., Parnes, P. & Milner, M. (1992), New Directions in Visual Keyboards for Graphical User Interfaces, in H. J. Murphy (ed.), “Proceedings of the 7th Annual Conference on Technology and Persons with Disabilities”, California State University, pp. 465–469.
Starker, I. & Bolt, R. A. (1990), A Gaze-Responsive Self-Disclosing Display, in J. C. Chew and J. Whiteside (eds.), “Proceedings of CHI’90: Human Factors in Computing Systems”, ACM Press, pp. 3–9.
Swiffin, A. L. (1988), A Predictive and Adaptive Communication System for the Handicapped, Master’s thesis, University of Dundee.
Ten Kate, J. H., Frietman, E. E. E., Stoel, F. J. M. L. & Willems, W. (1980), “Eye-Controlled Communication Aids”, Medical Progress through Technology 8, 1–21.
White, K. P., Hutchinson, T. E. & Carley, J. M. (1993), “Spatially Dynamic Calibration of an Eye-Tracking System”, IEEE Transactions in Systems, Man and Cybernetics 3(4), 11621168.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1996 Springer-Verlag London
About this paper
Cite this paper
Istance, H., Spinner, C., Howarth, P.A. (1996). Eye-based Control of Standard GUI Software. In: Sasse, M.A., Cunningham, R.J., Winder, R.L. (eds) People and Computers XI. Springer, London. https://doi.org/10.1007/978-1-4471-3588-3_10
Download citation
DOI: https://doi.org/10.1007/978-1-4471-3588-3_10
Publisher Name: Springer, London
Print ISBN: 978-3-540-76069-6
Online ISBN: 978-1-4471-3588-3
eBook Packages: Springer Book Archive