Abstract
This chapter presents an innovative semi-autonomous human-robot interaction concept for people with disability and discusses a proof-of-concept prototype system. The communication between the user and the robot is performed by electromyographic (EMG) signals. However, unlike most EMG controlled robotic operations, in this framework the user can issue high-level commands to the robot through a novel EMG based approach. The robot controller is designed such that it is capable of decomposing these high-level commands into primitive subtasks using task grammar. It then develops a dynamic plan that links these primitives with knowledge about the world view to accomplish the high-level task commands. In this manner, a user can achieve semi-autonomous human-robot interaction. This proposed concept eliminates the need for continuous control of the robot and as a result, makes the system easier to use, less tiring and less error-prone. This system provides a platform for the people with disability to supervise a robot through high-level qualitative commands, rather than through low-level robotic teleoperation directives. Such a system would permit a person with disability and a robot to communicate task-relevant information in a convenient, robust, and reliable manner.
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Editor information
Rights and permissions
About this chapter
Cite this chapter
Rani, P., Sarkar, M., Brackin, R., Sarkar, N. 8 Semi-Autonomous Human-Robot Interaction for People with Disability. In: Barbagli, F., Prattichizzo, D., Salisbury, K. (eds) Multi-point Interaction with Real and Virtual Objects. Springer Tracts in Advanced Robotics, vol 18. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11429555_8
Download citation
DOI: https://doi.org/10.1007/11429555_8
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-26036-3
Online ISBN: 978-3-540-31503-2
eBook Packages: EngineeringEngineering (R0)