Abstract
A Sign Language Recognition (SLR) system translates signs performed by deaf individuals into text/speech in real time. Low cost sensor modalities, inertial measurement unit (IMU) and surface electromyography (sEMG), are both useful to detect hand/arm gestures. They are capable of capturing signs and are complementary to each other for recognizing signs. In this book chapter, we propose a wearable system for recognizing American Sign Language (ASL) in real-time, fusing information from an inertial sensor and sEMG sensors. The best subset of features from a wide range of well-studied features is selected using an information gain based feature selection approach. Four popular classification algorithms are evaluated for 80 commonly used ASL signs on four subjects. With the selected feature subset and a support vector machine classifier, our system achieves 96.16 and 85.24% average accuracies for intra-subject and intra-subject cross session evaluation respectively. The significance of adding sEMG for American Sign Language recognition is explored and the best channel of sEMG is highlighted.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
W. C. Stokoe, “Sign language structure: An outline of the visual communication systems of the American deaf,” Journal of deaf studies and deaf education, vol. 10, no. 1, pp. 3–37, 2005.
D. Barberis, N. Garazzino, P. Prinetto, G. Tiotto, A. Savino, U. Shoaib, and N. Ahmad, “Language resources for computer assisted translation from italian to italian sign language of deaf people,” in Proceedings of Accessibility Reaching Everywhere AEGIS Workshop and International Conference, Brussels, Belgium (November 2011), 2011.
A. B. Grieve-Smith, “Signsynth: A sign language synthesis application using web3d and PERL,” in Gesture and Sign Language in Human-Computer Interaction, pp. 134–145, Springer, 2002.
B. Vicars, “Basic ASL: First 100 signs.”
E. Costello, American sign language dictionary. Random House Reference &, 2008.
T. Starner, J. Weaver, and A. Pentland, “Real-time american sign language recognition using desk and wearable computer based video,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 20, no. 12, pp. 1371–1375, 1998.
C. Vogler and D. Metaxas, “A framework for recognizing the simultaneous aspects of American sign language,” Computer Vision and Image Understanding, vol. 81, no. 3, pp. 358–384, 2001.
T. E. Starner, “Visual recognition of American sign language using Hidden Markov models.,” tech. rep., DTIC Document, 1995.
C. Oz and M. C. Leu, “American sign language word recognition with a sensory glove using artificial neural networks,” Engineering Applications of Artificial Intelligence, vol. 24, no. 7, pp. 1204–1213, 2011.
E. Malaia, J. Borneman, and R. B. Wilbur, “Analysis of ASL motion capture data towards identification of verb type,” in Proceedings of the 2008 Conference on Semantics in Text Processing, pp. 155–164, Association for Computational Linguistics, 2008.
A. Y. Benbasat and J. A. Paradiso, “An inertial measurement framework for gesture recognition and applications,” in Gesture and Sign Language in Human-Computer Interaction, pp. 9–20, Springer, 2002.
O. Amft, H. Junker, and G. Troster, “Detection of eating and drinking arm gestures using inertial body-worn sensors,” in Wearable Computers, 2005. Proceedings. Ninth IEEE International Symposium on, pp. 160–163, IEEE, 2005.
A. B. Ajiboye and R. F. Weir, “A heuristic fuzzy logic approach to EMG pattern recognition for multifunctional prosthesis control,” Neural Systems and Rehabilitation Engineering, IEEE Transactions on, vol. 13, no. 3, pp. 280–291, 2005.
J.-U. Chu, I. Moon, and M.-S. Mun, “A real-time EMG pattern recognition based on linear-nonlinear feature projection for multifunction myoelectric hand,” in Rehabilitation Robotics, 2005. ICORR 2005. 9th International Conference on, pp. 295–298, IEEE, 2005.
Y. Li, X. Chen, X. Zhang, K. Wang, and J. Yang, “Interpreting sign components from accelerometer and sEMG data for automatic sign language recognition,” in Engineering in Medicine and Biology Society, EMBC, 2011 Annual International Conference of the IEEE, pp. 3358–3361, IEEE, 2011.
Y. Li, X. Chen, X. Zhang, K. Wang, and Z. J. Wang, “A sign-component-based framework for chinese sign language recognition using accelerometer and sEMG data,” Biomedical Engineering, IEEE Transactions on, vol. 59, no. 10, pp. 2695–2704, 2012.
L.-G. Zhang, Y. Chen, G. Fang, X. Chen, and W. Gao, “A vision-based sign language recognition system using tied-mixture density hmm,” in Proceedings of the 6th international conference on Multimodal interfaces, pp. 198–204, ACM, 2004.
M. M. Zaki and S. I. Shaheen, “Sign language recognition using a combination of new vision based features,” Pattern Recognition Letters, vol. 32, no. 4, pp. 572–577, 2011.
T.-Y. Pan, L.-Y. Lo, C.-W. Yeh, J.-W. Li, H.-T. Liu, and M.-C. Hu, “Real-time sign language recognition in complex background scene based on a hierarchical clustering classification method,” in Multimedia Big Data (BigMM), 2016 IEEE Second International Conference on, pp. 64–67, IEEE, 2016.
M. W. Kadous et al., “Machine recognition of AUSLAN signs using powergloves: Towards large-Lexicon recognition of sign language,” in Proceedings of the Workshop on the Integration of Gesture in Language and Speech, pp. 165–174, Citeseer, 1996.
M. G. Kumar, M. K. Gurjar, and M. S. B. Singh, “American sign language translating glove using flex sensor,” Imperial Journal of Interdisciplinary Research, vol. 2, no. 6, 2016.
D. Sherrill, P. Bonato, and C. De Luca, “A neural network approach to monitor motor activities,” in Engineering in Medicine and Biology, 2002. 24th Annual Conference and the Annual Fall Meeting of the Biomedical Engineering Society EMBS/BMES Conference, 2002. Proceedings of the Second Joint, vol. 1, pp. 52–53, IEEE, 2002.
X. Chen, X. Zhang, Z.-Y. Zhao, J.-H. Yang, V. Lantz, and K.-Q. Wang, “Hand gesture recognition research based on surface EMG sensors and 2D-accelerometers,” in Wearable Computers, 2007 11th IEEE International Symposium on, pp. 11–14, IEEE, 2007.
V. E. Kosmidou and L. J. Hadjileontiadis, “Sign language recognition using intrinsic-mode sample entropy on sEMG and accelerometer data,” Biomedical Engineering, IEEE Transactions on, vol. 56, no. 12, pp. 2879–2890, 2009.
S. Wei, X. Chen, X. Yang, S. Cao, and X. Zhang, “A component-based vocabulary-extensible sign language gesture recognition framework,” Sensors, vol. 16, no. 4, p. 556, 2016.
J. Kim, J. Wagner, M. Rehm, and E. André, “Bi-channel sensor fusion for automatic sign language recognition,” in Automatic Face & Gesture Recognition, 2008. FG’08. 8th IEEE International Conference on, pp. 1–6, IEEE, 2008.
J.-S. Wang and F.-C. Chuang, “An accelerometer-based digital pen with a trajectory recognition algorithm for handwritten digit and gesture recognition,” Industrial Electronics, IEEE Transactions on, vol. 59, no. 7, pp. 2998–3007, 2012.
V. Nathan, J. Wu, C. Zong, Y. Zou, O. Dehzangi, M. Reagor, and R. Jafari, “A 16-channel bluetooth enabled wearable EEG platform with dry-contact electrodes for brain computer interface,” in Proceedings of the 4th Conference on Wireless Health, p. 17, ACM, 2013.
C. J. De Luca, L. Donald Gilmore, M. Kuznetsov, and S. H. Roy, “Filtering the surface emg signal: Movement artifact and baseline noise contamination,” Journal of biomechanics, vol. 43, no. 8, pp. 1573–1579, 2010.
I. Mesa, A. Rubio, I. Tubia, J. De No, and J. Diaz, “Channel and feature selection for a surface electromyographic pattern recognition task,” Expert Systems with Applications, vol. 41, no. 11, pp. 5190–5200, 2014.
R. Merletti and P. Di Torino, “Standards for reporting EMG data,” J Electromyogr Kinesiol, vol. 9, no. 1, pp. 3–4, 1999.
A. Phinyomark, C. Limsakul, and P. Phukpattaranont, “A novel feature extraction for robust EMG pattern recognition,” arXiv preprint arXiv:0912.3973, 2009.
M. Zhang and A. A. Sawchuk, “Human daily activity recognition with sparse representation using wearable sensors,” Biomedical and Health Informatics, IEEE Journal of, vol. 17, no. 3, pp. 553–560, 2013.
S. H. Khan and M. Sohail, “Activity monitoring of workers using single wearable inertial sensor.”
O. Paiss and G. F. Inbar, “Autoregressive modeling of surface EMG and its spectrum with application to fatigue,” Biomedical Engineering, IEEE Transactions on, no. 10, pp. 761–770, 1987.
A. M. Khan, Y.-K. Lee, S. Y. Lee, and T.-S. Kim, “A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer,” Information Technology in Biomedicine, IEEE Transactions on, vol. 14, no. 5, pp. 1166–1172, 2010.
I. Guyon and A. Elisseeff, “An introduction to variable and feature selection,” The Journal of Machine Learning Research, vol. 3, pp. 1157–1182, 2003.
J. R. Quinlan, C4. 5: programs for machine learning. Elsevier, 2014.
C.-C. Chang and C.-J. Lin, “LIBSVM: A library for support vector machines,” ACM Transactions on Intelligent Systems and Technology, vol. 2, pp. 27:1–27:27, 2011. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvm.
M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten, “The weka data mining software: an update,” ACM SIGKDD explorations newsletter, vol. 11, no. 1, pp. 10–18, 2009.
M. Z. Jamal, “Signal acquisition using surface EMG and circuit design considerations for robotic prosthesis,” 2012.
J. Wu, Z. Tian, L. Sun, L. Estevez, and R. Jafari, “Real-time american sign language recognition using wrist-worn motion and surface EMG sensors,” in Wearable and Implantable Body Sensor Networks (BSN), 2015 IEEE 12th International Conference on, pp. 1–6, IEEE, 2015.
Acknowledgements
This work was supported in part by the National Science Foundation, under grants CNS-1150079 and ECCS-1509063. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the funding organizations.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this chapter
Cite this chapter
Wu, J., Jafari, R. (2017). Wearable Computers for Sign Language Recognition. In: Khan, S., Zomaya, A., Abbas, A. (eds) Handbook of Large-Scale Distributed Computing in Smart Healthcare. Scalable Computing and Communications. Springer, Cham. https://doi.org/10.1007/978-3-319-58280-1_14
Download citation
DOI: https://doi.org/10.1007/978-3-319-58280-1_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-58279-5
Online ISBN: 978-3-319-58280-1
eBook Packages: Computer ScienceComputer Science (R0)