skip to main content
research-article

SmartASL: "Point-of-Care" Comprehensive ASL Interpreter Using Wearables

Authors Info & Claims
Published:12 June 2023Publication History
Skip Abstract Section

Abstract

Sign language builds up an important bridge between the d/Deaf and hard-of-hearing (DHH) and hearing people. Regrettably, most hearing people face challenges in comprehending sign language, necessitating sign language translation. However, state-of-the-art wearable-based techniques mainly concentrate on recognizing manual markers (e.g., hand gestures), while frequently overlooking non-manual markers, such as negative head shaking, question markers, and mouthing. This oversight results in the loss of substantial grammatical and semantic information in sign language. To address this limitation, we introduce SmartASL, a novel proof-of-concept system that can 1) recognize both manual and non-manual markers simultaneously using a combination of earbuds and a wrist-worn IMU, and 2) translate the recognized American Sign Language (ASL) glosses into spoken language. Our experiments demonstrate the SmartASL system's significant potential to accurately recognize the manual and non-manual markers in ASL, effectively bridging the communication gaps between ASL signers and hearing people using commercially available devices.

References

  1. Accessibility.com, LLC. 2022. Is American Sign Language (ASL) a language? https://www.accessibility.com/blog/is-american-sign-language-asl-a-language/.Google ScholarGoogle Scholar
  2. Ashwin Ahuja, Andrea Ferlini, and Cecilia Mascolo. 2021. PilotEar: Enabling In-ear Inertial Navigation. In Adjunct Proceedings of the 2021 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2021 ACM International Symposium on Wearable Computers. 139--145.Google ScholarGoogle Scholar
  3. Takashi Amesaka, Hiroki Watanabe, and Masanori Sugimoto. 2019. Facial expression recognition using ear canal transfer function. In Proceedings of the 23rd International Symposium on Wearable Computers. 1--9.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. F Berzin and CRH Fortinguerra. 1993. EMG study of the anterior, superior and posterior auricular muscles in man. Annals of Anatomy-Anatomischer Anzeiger 175, 2 (1993), 195--197.Google ScholarGoogle ScholarCross RefCross Ref
  5. Hongliang Bi and Jiajia Liu. 2022. CSEar: Meta-learning for Head Gesture Recognition Using Earphones in Internet of Healthcare Things. IEEE Internet of Things Journal (2022).Google ScholarGoogle Scholar
  6. Eric Branda and Tobias Wurzbacher. 2021. Motion Sensors in Automatic Steering of Hearing Aids. In Seminars in Hearing, Vol. 42. Thieme Medical Publishers, Inc., 237--247.Google ScholarGoogle Scholar
  7. Nam Bui, Nhat Pham, Jessica Jacqueline Barnitz, Zhanan Zou, Phuc Nguyen, Hoang Truong, Taeho Kim, Nicholas Farrow, Anh Nguyen, Jianliang Xiao, et al. 2019. ebp: A wearable system for frequent and comfortable blood pressure monitoring from user's ear. In The 25th Annual International Conference on Mobile Computing and Networking. 1--17.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Kayla-Jade Butkow, Ting Dang, Andrea Ferlini, Dong Ma, and Cecilia Mascolo. 2021. Motion-resilient Heart Rate Monitoring with In-ear Microphones. arXiv preprint arXiv:2108.09393 (2021).Google ScholarGoogle Scholar
  9. George Caridakis, Stylianos Asteriadis, and Kostas Karpouzis. 2014. Non-manual cues in automatic sign language recognition. Personal and ubiquitous computing 18, 1 (2014), 37--46.Google ScholarGoogle Scholar
  10. Seokmin Choi, Yang Gao, Yincheng Jin, Se jun Kim, Jiyang Li, Wenyao Xu, and Zhanpeng Jin. 2022. PPGface: Like What You Are Watching? Earphones Can" Feel" Your Facial Expressions. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, 2 (2022), 1--32.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Deaf Community. 2021. Deaf Culture. https://www.startasl.com/what-does-d-d-and-d-deaf-mean-in-the-deaf-community/. [Updated May 13, 2021].Google ScholarGoogle Scholar
  12. ASLLRP Continuous Signing Corpora. 2022. American Sign Language Linguistic Research Project. https://dai.cs.rutgers.edu/dai/s/dai. [Online].Google ScholarGoogle Scholar
  13. Biyi Fang, Jillian Co, and Mi Zhang. 2017. DeepASL: Enabling ubiquitous and non-intrusive word and sentence-level sign language translation. In Proceedings of the 15th ACM Conference on Embedded Network Sensor Systems. 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Andrea Ferlini, Dong Ma, Robert Harle, and Cecilia Mascolo. 2021. EarGate: gait-based user identification with in-ear microphones. In Proceedings of the 27th Annual International Conference on Mobile Computing and Networking. 337--349.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. E Friesen and Paul Ekman. 1978. Facial action coding system: a technique for the measurement of facial movement. Palo Alto 3, 2 (1978), 5.Google ScholarGoogle Scholar
  16. Yang Gao, Yincheng Jin, Seokmin Choi, Jiyang Li, Junjie Pan, Lin Shu, Chi Zhou, and Zhanpeng Jin. 2021. SonicFace: Tracking Facial Expressions Using a Commodity Microphone Array. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 4 (2021), 1--33.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Yang Gao, Wei Wang, Vir V. Phoha, Wei Sun, and Zhanpeng Jin. 2019. EarEcho: Using Ear Canal Echo for Wearable Authentication. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3, 3, Article 81 (Sept. 2019), 24 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Google. 2022. AR Glass. https://nerdist.com/article/google-ar-glasses-live-translation-real-time-transcription/.Google ScholarGoogle Scholar
  19. Audien Hearing. 2023. Atom Pro. https://audienhearing.com/products/audien-atom-pro-pair?variant=39511193255999.Google ScholarGoogle Scholar
  20. Carl-Herman Hjortsjö. 1969. Man's face and mimic language. Studentlitteratur.Google ScholarGoogle Scholar
  21. Jiahui Hou, Xiang-Yang Li, Peide Zhu, Zefan Wang, Yu Wang, Jianwei Qian, and Panlong Yang. 2019. SignSpeaker: A real-time, high-precision smartwatch-based sign language translator. In The 25th Annual International Conference on Mobile Computing and Networking (MobiCom'19). Article 24, 15 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Jie Huang, Wengang Zhou, Qilin Zhang, Houqiang Li, and Weiping Li. 2018. Video-based sign language recognition without temporal segmentation. In Thirty-Second AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  23. Yincheng Jin, Yang Gao, Xiaotao Guo, Jun Wen, Zhengxiong Li, and Zhanpeng Jin. 2022. EarHealth: an earphone-based acoustic otoscope for detection of multiple ear diseases in daily life. In Proceedings of the 20th Annual International Conference on Mobile Systems, Applications and Services. 397--408.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Yincheng Jin, Yang Gao, Xuhai Xu, Seokmin Choi, Jiyang Li, Feng Liu, Zhengxiong Li, and Zhanpeng Jin. 2022. EarCommand: "Hearing" Your Silent Speech Commands In Ear. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, 2 (2022), 1--28.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Yincheng Jin, Yang Gao, Yanjun Zhu, Wei Wang, Jiyang Li, Seokmin Choi, Zhangyu Li, Jagmohan Chauhan, Anind K Dey, and Zhanpeng Jin. 2021. SonicASL: An acoustic-based sign language gesture recognizer using earphones. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 2 (2021), 1--30.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Sara Askari Khomami and Sina Shamekhi. 2021. Persian sign language recognition using IMU and surface EMG sensors. Measurement 168 (2021), 108471.Google ScholarGoogle ScholarCross RefCross Ref
  27. Suyoun Kim, Takaaki Hori, and Shinji Watanabe. 2017. Joint CTC-attention based end-to-end speech recognition using multi-task learning. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 4835--4839.Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Nicolas Le Goff, Jesper Jensen, Michael Syskind Pedersen, and Susanna Løve Callaway. 2016. An introduction to OpenSound Navigator™. Oticon A/S (2016).Google ScholarGoogle Scholar
  29. Steven F LeBoeuf, Michael E Aumer, William E Kraus, Johanna L Johnson, and Brian Duscha. 2014. Earbud-based sensor for the assessment of energy expenditure, heart rate, and VO2max. Medicine and Science in Sports and Exercise 46, 5 (2014), 1046.Google ScholarGoogle ScholarCross RefCross Ref
  30. Yilin Liu, Fengyang Jiang, and Mahanth Gowda. 2020. Finger Gesture Tracking for Interactive Applications: A Pilot Study with Sign Languages. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 4, 3 (2020), 1--21.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Yilin Liu, Shijia Zhang, and Mahanth Gowda. 2021. When Video Meets Inertial Sensors: Zero-Shot Domain Adaptation for Finger Motion Analytics with Inertial Sensors. In Proceedings of the International Conference on Internet-of-Things Design and Implementation (Charlottesvle, VA, USA) (IoTDI '21). ACM, New York, NY, USA, 182--194.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Hamzah Luqman and El-Sayed M El-Alfy. 2021. Towards hybrid multimodal manual and non-manual Arabic sign language recognition: MArSL database and pilot study. Electronics 10, 14 (2021), 1739.Google ScholarGoogle ScholarCross RefCross Ref
  33. Yongsen Ma, Gang Zhou, Shuangquan Wang, Hongyang Zhao, and Woosub Jung. 2018. SignFi: Sign language recognition using WiFi. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 2, 1, Article 23 (2018), 21 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Stephen McCullough, Karen Emmorey, and Martin Sereno. 2005. Neural organization for recognition of grammatical and emotional facial expressions in deaf ASL signers and hearing nonsigners. Cognitive Brain Research 22, 2 (2005), 193--203.Google ScholarGoogle ScholarCross RefCross Ref
  35. Meta. 2016. Binaural Audio for Narrative AR. https://www.oculus.com/story-studio/blog/binaural-audio-for-narrative-vr/.Google ScholarGoogle Scholar
  36. Nicholas Michael, Peng Yang, Qingshan Liu, Dimitris N Metaxas, Carol Neidle, and CBIM Center. 2011. A Framework for the Recognition of Nonmanual Markers in Segmented Sequences of American Sign Language.. In BMVC. 1--12.Google ScholarGoogle Scholar
  37. NIH. 2008. Hearing Loss and Hearing Aid Use. https://www.mdcd.nih.gov/news/multimedia/hearing-loss-and-hearing-aid-use-text-version. [Updated July 17, 2017].Google ScholarGoogle Scholar
  38. Achraf Othman and Mohamed Jemni. 2012. English-ASL gloss parallel corpus 2012: ASLG-PC12. In 5th Workshop on the Representation and Processing of Sign Languages: Interactions between Corpus and Lexicon LREC.Google ScholarGoogle Scholar
  39. Sinno Jialin Pan and Qiang Yang. 2009. A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering 22, 10 (2009), 1345--1359.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J Liu, et al. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 140 (2020), 1--67.Google ScholarGoogle Scholar
  41. Grand Review Research. 2023. Grand Review Research. https://www.grandviewresearch.com/industry-analysis/earphone-and-headphone-market. [Online].Google ScholarGoogle Scholar
  42. Tobias Röddiger, Christopher Clarke, Paula Breitling, Tim Schneegans, Haibin Zhao, Hans Gellersen, and Michael Beigl. 2022. Sensing with Earables: A Systematic Literature Review and Taxonomy of Phenomena. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, 3 (2022), 1--57.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Arman Sabyrov, Medet Mukushev, and Vadim Kimmelman. 2019. Towards Real-time Sign Language Interpreting Robot: Evaluation of Non-manual Components on Recognition Accuracy.. In CVPR Workshops.Google ScholarGoogle Scholar
  44. Panneer Selvam Santhalingam, Al Amin Hosain, Ding Zhang, Parth Pathak, Huzefa Rangwala, and Raja Kushalnagar. 2020. mmASL: Environment-Independent ASL Gesture Recognition Using 60 GHz Millimeter-wave Signals. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 4, 1, Article 26 (2020), 30 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Torgyn Shaikhina and Natalia A. Khovanova. 2017. Handling limited datasets with neural networks in medical applications: A small-data approach. Artificial Intelligence in Medicine 75 (2017), 51--63.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Jiacheng Shang and Jie Wu. 2017. A robust sign language recognition system with multiple Wi-Fi devices. In Proceedings of the Workshop on Mobility in the Evolving Internet Architecture. 19--24.Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. Xingzhe Song, Kai Huang, and Wei Gao. 2022. FaceListener: Recognizing Human Facial Expressions via Acoustic Sensing on Commodity Headphones. In 2022 21st ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN). IEEE, 145--157.Google ScholarGoogle ScholarCross RefCross Ref
  48. StartASL. 2020. ASL Dictionary -- Learn Essential Vocabulary. https://www.handspeak.com/word/. [Updated April 28, 2020].Google ScholarGoogle Scholar
  49. Karush Suri and Rinki Gupta. 2019. Continuous sign language recognition from wearable IMUs using deep capsule networks and game theory. Computers & Electrical Engineering 78 (2019), 493--503.Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. Noeru Suzuki, Yuki Watanabe, and Atsushi Nakazawa. 2020. Gan-based style transformation to improve gesture-recognition accuracy. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (IMWUT) 4, 4 (2020), 1--20.Google ScholarGoogle ScholarDigital LibraryDigital Library
  51. Andrius Vabalas, Emma Gowen, Ellen Poliakoff, and Alexander J. Casson. 2019. Machine learning algorithm validation with a limited sample size. PLoS ONE 14, 11 (2019), 1--20.Google ScholarGoogle ScholarCross RefCross Ref
  52. Dhruv Verma, Sejal Bhalla, Dhruv Sahnan, Jainendra Shukla, and Aman Parnami. 2021. ExpressEar: Sensing Fine-Grained Facial Expressions with Earables. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 3 (2021), 1--28.Google ScholarGoogle ScholarDigital LibraryDigital Library
  53. Zi Wang, Sheng Tan, Linghan Zhang, Yili Ren, Zhi Wang, and Jie Yang. 2021. EarDynamic: An Ear Canal Deformation Based Continuous User Authentication Using In-Ear Wearables. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 1 (2021), 1--27.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. Katharine L Watson. 2010. WH-questions in American Sign Language: Contributions of non-manual marking to structure and meaning. Purdue University.Google ScholarGoogle Scholar
  55. Traci Patricia Weast. 2008. Questions in American Sign Language: A quantitative analysis of raised and lowered eyebrows. The University of Texas at Arlington.Google ScholarGoogle Scholar
  56. WHO. 2022. Deafness and hearing loss. https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss. [Online].Google ScholarGoogle Scholar
  57. Jian Wu, Lu Sun, and Roozbeh Jafari. 2016. A Wearable System for Recognizing American Sign Language in Real-Time Using IMU and Surface EMG Sensors. IEEE Journal of Biomedical and Health Informatics 20, 5 (2016), 1281--1290.Google ScholarGoogle ScholarCross RefCross Ref
  58. Kayo Yin. 2020. Sign language translation with transformers. arXiv preprint arXiv:2004.00588 2 (2020).Google ScholarGoogle Scholar
  59. Zahoor Zafrulla, Helene Brashear, Thad Starner, Harley Hamilton, and Peter Presti. 2011. American sign language recognition with the kinect. In Proceedings of the 13th International Conference on Multimodal Interfaces. 279--286.Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. Qian Zhang, JiaZhen Jing, Dong Wang, and Run Zhao. 2022. WearSign: Pushing the Limit of Sign Language Translation Using Inertial and EMG Wearables. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, 1 (2022), 1--27.Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. Qian Zhang, Dong Wang, Run Zhao, and Yinggang Yu. 2019. MyoSign: enabling end-to-end sign language recognition with wearables. In Proceedings of the 24th International Conference on Intelligent User Interfaces. 650--660.Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. Zhihao Zhou, Kyle Chen, Xiaoshi Li, Songlin Zhang, Yufen Wu, Yihao Zhou, Keyu Meng, Chenchen Sun, Qiang He, Wenjing Fan, Endong Fan, Zhiwei Lin, Xulong Tan, Weili Deng, Jin Yang, and Jun Chen. 2020. Sign-to-speech translation using machine-learning-assisted stretchable sensor arrays. Nature Electronics 3 (2020), 571--578.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. SmartASL: "Point-of-Care" Comprehensive ASL Interpreter Using Wearables

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
      Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies  Volume 7, Issue 2
      June 2023
      969 pages
      EISSN:2474-9567
      DOI:10.1145/3604631
      Issue’s Table of Contents

      Copyright © 2023 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 12 June 2023
      Published in imwut Volume 7, Issue 2

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader