Abstract
This paper presents a novel human machine interface for people with severe disabilities to control an electric powered wheelchair using face movements. Five face movements including jaw clenching and eye closing movements are identified by extracting movement features from both forehead Electromyography (EMG) signal and facial image information. A real-world indoor environment is setup for evaluating the performance of the new control method. Five subjects participated in the experiment to follow designated routes on a map using the new control method, as well as a traditional joystick control respectively. Comparison of two control methods are made in terms of easiness of control, time duration, wheelchair trajectory and error command rate etc. Wheelchair trajectory and time consumption are recorded for each task and results show the new control method are comparable to a joystick control and can be used as a hands-free control interface for disabled and elderly users.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Felzer, T., Freisleben, B.: HaWCoS: The ‘Hands-free’ Wheelchair Control System. In: Proceedings of International ACM SIGACCESS Conference on Computers and Accessibility, pp. 127–134. ACM Press, New York (2002)
Moon, I., Lee, M., Chu, J., Mun, M.: Wearable EMG-based HCI for electric-powered wheelchair users with motor disabilities. In: Proceedings of IEEE International Conference on Robotics and Automation, pp. 2649–2654 (2005)
Barea, R., Boquete, L., Mazo, M., Lopez, E.: System for assisted mobility using eye movements based on electrooculography. IEEE Transactions on Neural Systems and Rehabilitation Engineering 10, 209–218 (2002)
Rebsamen, B., Burdet, E., Guan, C., Zhang, H., Teo, C.L., Zeng, Q., Laugier Jr., C., Marcelo, A.H.: Controlling a Wheelchair Indoors Using Thought. IEEE Intelligent Systems 22(2), 18–24 (2007)
Tsui, C.S.L., Jia, P., Gan, J.Q., Hu, H., Yuan, K.: EMG-based hands-free wheelchair control with EOG attention shift detection. In: Proceedings of IEEE International Conference on Robotics and Biomimetics, pp. 1266–1271 (December 2007)
Ju, J.S., Shin, Y., Kim, E.Y.: Vision based interface system for hands free control of an Intelligent Wheelchair. Journal of Neuroengineering and Rehabilitation 6 (2009)
Li, X., Tan, T., Zhao, X.: Multi-modal Navigation for Interactive Wheelchair. In: Tan, T., Shi, Y., Gao, W. (eds.) ICMI 2000. LNCS, vol. 1948, pp. 590–598. Springer, Heidelberg (2000)
Matsumoto, Y., Ino, T., Ogasawara, T.: Development of Intelligent Wheelchair System with Face and Gaze Based Interface. In: Proceedings of 10th IEEE Int. Workshop on Robot and Human Communication (ROMAN 2001), pp. 262–267 (2001)
Ferreira, A., Silva, R.L., Celeste, W.C., Bastos Filho, T.F., Sarcinelli Filho, M.: Human–machine interface based on muscular and brain signals applied to a robotic wheelchair. Journal of Physics: Conference Series 90, 012094 (2007)
Cyberlink Brainfingers, Hands-free Computer Access Solution, http://www.brainfingers.com/
Jia, P., Hu, H., Lu, T., Yuan, K.: Head gesture recognition for hands-free control of an intelligent wheelchair. Journal of Industrial Robot 34(1), 60–68 (2007)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Wei, L., Hu, H. (2010). Multi-modality — EMG and Visual Based Hands-Free Control of an Intelligent Wheelchair. In: Liu, H., Ding, H., Xiong, Z., Zhu, X. (eds) Intelligent Robotics and Applications. ICIRA 2010. Lecture Notes in Computer Science(), vol 6425. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-16587-0_60
Download citation
DOI: https://doi.org/10.1007/978-3-642-16587-0_60
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-16586-3
Online ISBN: 978-3-642-16587-0
eBook Packages: Computer ScienceComputer Science (R0)