Abstract
Efforts to make online media accessible to a regional audience have picked up pace in recent years with multilingual captioning and keyboards. However, techniques to extend this access to people with hearing loss are limited. Further, owing to a lack of structure in the education of hearing impaired and regional differences, the issue of standardization of Indian Sign Language (ISL) has been left unaddressed, forcing educators to rely on the local language to support the ISL structure, thereby creating an array of correlations for each object, hindering the language building skills of a student. This paper aims to present a useful technology that can be used to leverage online resources and make them accessible to the hearing-impaired community in their primary mode of communication. Our tool presents an avenue for the early development of language learning and communication skills essential for the education of children with a profound hearing loss. With the proposed technology, we aim to provide a standardized teaching and learning medium to a classroom setting that can utilize and promote ISL. The goals of our proposed system involve reducing the burden of teachers to act as a valuable teaching aid. The system allows for easy translation of any online video and correlation with ISL captioning using a 3D cartoonish avatar aimed to reinforce classroom concepts during the critical period. First, the video gets converted to text via subtitles and speech processing methods. The generated text is understood through NLP algorithms and then mapped to avatar captions which are then rendered to form a cohesive video alongside the original content. We validated our results through a 6-month period and a consequent 2-month study, where we recorded a 37% and 70% increase in performance of students taught using Sign captioned videos against student taught with English captioned videos. We also recorded a 73.08% increase in vocabulary acquisition through signed aided videos.
Similar content being viewed by others
References
Agarwal, A., Thakur, M.K.: Sign language recognition using microsoft kinect. In: 2013 Sixth International Conference on Contemporary Computing (IC3), pp. 181–185 (2013). https://doi.org/10.1109/IC3.2013.6612186
Ahmed, M.A., Zaidan, B.B., Zaidan, A.A., Salih, M.M., Lakulu, M.M.B.: A review on systems-based sensory gloves for sign language recognition state of the art between 2007 and 2017. Sensors 18(7), 2208 (2018). https://doi.org/10.3390/s18072208
Ahmed, S.: Real time American sign language video captioning using deep neural networks [PowerPoint presentation]. http://on-demand.gputechconf.com/gtc/2017/presentation/s7346-syed-ahmed-real-time-american-sign-language-video-caption.pdf (2018)
Cannon, J.E., Fredrick, L.D., Easterbrooks, S.R.: Vocabulary instruction through books read in American sign language for English-language learners with hearing loss. Commun. Disord. Q. 31(2), 98–112 (2010). https://doi.org/10.1177/1525740109332832
Center, L.C.N.D.E.: Learning american sign language: Books, media, and classes. http://www3.gallaudet.edu/clerc-center/info-to-go/asl/learning-asl-books_media_classes.html. [Online]
Chambers, B., Abrami, P.C., McWhaw, K., Therrien, M.C.: Developing a computer-assisted tutoring program to help children at risk learn to read. Educ. Res. Eval. 7(2–3), 223–239 (2001). https://doi.org/10.1076/edre.7.2.223.3863
Chandramouli, C.: Census of India 2011, provisional population totals, government of India. http://censusindia.gov.in/2011-prov-results/paper2/data_files/india/paper2_1.pdf (2011). [Online]
Division, S.S.: Disabled persons in India, a statistical profile 2016. http://mospi.nic.in/sites/default/files/publication_reports/Disabled_persons_in_India_2016.pdf (2016). [Online]
Easterbrooks, S.R., Huston, S.G.: The signed reading fluency of students who are deaf/hard of hearing. J. Deaf Stud. Deaf Educ. 13(1), 37–54 (2007). https://doi.org/10.1093/deafed/enm030
Fung, P.C., Chow, B.W.Y., McBride-Chang, C.: The impact of a dialogic reading program on deaf and hard-of-hearing kindergarten and early primary school-aged students in Hong Kong. J. Deaf Educ. Deaf Stud. 10(1), 82–95 (2005)
Goyal, L., Goyal, V.: Automatic translation of English text to Indian sign language synthetic animations. In: Proceedings of the 13th International Conference on Natural Language Processing, pp. 144–153 (2016)
Hermans, D., Knoors, H., Ormel, E., Verhoeven, L.: The relationship between the reading and signing skills of deaf children in bilingual education programs. J. Deaf Stud. Deaf Educ. 13(4), 518–530 (2008). https://doi.org/10.1093/deafed/enn009
Huenerfauth, M.: Generating American sign language animation: overcoming misconceptions and technical challenges. Univers. Access Inf. Soc. 6(4), 419–434 (2008). https://doi.org/10.1007/s10209-007-0095-7
Journal, C.W.R.: Oralism and how it affects the development of the deaf child. http://www.csus.edu/wac/journal/2010/hood_research_paper.pdf (2010)
Kafle, S., Huenerfauth, M.: Evaluating the usability of automatically generated captions for people who are deaf or hard of hearing. In: Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS’17, pp. 165–174. ACM, New York, NY, USA (2017). https://doi.org/10.1145/3132525.3132542
Lang, S., Block, M., Rojas, R.: Sign language recognition using kinect. In: Rutkowski, L., Korytkowski, M., Scherer, R., Tadeusiewicz, R., Zadeh, L.A., Zurada, J.M. (eds.) Artificial Intelligence and Soft Computing, pp. 394–402. Springer Berlin Heidelberg, Berlin (2012)
Lazar, J., Feng, J.H., Hochheiser, H.: Research Methods in Human–Computer Interaction. Wiley, New York (2010)
Justice, L.M.: Word exposure conditions and preschoolers’ novel word learning during shared storybook reading. Read. Psychol. 23(2), 87–106 (2002). https://doi.org/10.1080/027027102760351016
Massaro, D.W., Light, J.: Improving the vocabulary of children with hearing loss. Volta Rev. 104(3), 141–174 (2004)
Mayberry, R.I., Eichen, E.B.: The long-lasting advantage of learning sign language in childhood: another look at the critical period for language acquisition. J. Mem. Lang. 30(4), 486–512 (1991). https://doi.org/10.1016/0749-596X(91)90018-F
Michael Strong, P.M.P.: A study of the relationship between American sign language and English literacy. J. Deaf Stud. Deaf Educ. 2(1), 37–46 (1997)
NCERT: Chapter-wise video lessons (2018). http://www.ncert.nic.in/new_ncert/ncert/publication/publication_list/list_of_publication3.html. [Online]
Capirci, O., Cattani, A., Rossini, P., Volterra, V.: Teaching sign language to hearing children as a possible factor in cognitive enhancement. J. Deaf Stud. Deaf Educ. 3(2), 135–142 (1998). https://doi.org/10.1093/oxfordjournals.deafed.a014343
Pettengill, R.: Motionsavvy uni (2011). http://www.motionsavvy.com/about.html. [Online]
RKMVERI: Isl dictionary (2004). http://www.indiansignlanguage.org/. [Online]
Robbins, C., Ehri, L.C.: Reading storybooks to kindergartners helps them learn new vocabulary words. J. Educ. Psychol. 86(1), 54–64 (1994). https://doi.org/10.1037/0022-0663.86.1.54
Santorini, B.: Part-Of-Speech tagging guidelines for the Penn Treebank project (3rd revision, 2nd printing). Tech. rep., Department of Linguistics, University of Pennsylvania, Philadelphia, PA, USA (1990)
Schleper, D.R.: Read it again... and again and again. Perspect. Educ. Deaf. 14(2), 16–19 (1995)
Society, D.A.: Sheila Kothavala Institute for the Deaf. http://deafaidsociety.in/ (2018)
Sousa, L., Rodrigues, J.M.F., Monteiro, J., Cardoso, P.J.S., Lam, R.: Gygsla: a portable glove system for learning sign language alphabet. In: Antona, M., Stephanidis, C. (eds.) Universal Access in Human-Computer Interaction. Users and Context Diversity, pp. 159–170. Springer, Cham (2016)
Strong, M., Prinz, P.: A study of the relationship between American sign language and English literacy. J. Deaf Stud. Deaf Educ. 2(1), 37–46 (1997)
Talk, S.: Signtalk: The resource for sign language interpreters. http://signtalk.org/ (2018)
Talkinghands: Indian sign language dictionary (2013). http://www.talkinghands.co.in/. [Online]
Tenório, R.: Virtual interpreter for Brazillian sign language. https://www.handtalk.me/ (2012). [Online]
Verma, D., Dash, P., Bhaskar, S., Pal, R.P., Jain, K., Srivastava, R.P., Hansraj, N.: Disabled persons in India: a statistical profile 2016. http://mospi.nic.in/sites/default/files/publication_reports/Disabled_persons_in_India_2016.pdf (2017). [Online]
Karnabadhir Vidyalaya, P.: PADSAD-the unique school aiming to main stream every hearing impairment and divyang student. http://www.padsad.org/ (2018)
Wideman, C.: Vcom3d. http://www.vcom3d.com/ (2002)
Wiggins, E.: Total communication as an education philosophy. http://www.deaflinx.com/DeafEd/OptionsGuide/TC.html (1998). [Online]
Wobbrock, J.O., Kientz, J.A.: Research contributions in human–computer interaction. Interactions 23(3), 38–44 (2016). https://doi.org/10.1145/2907069
Ye, Y., Tian, Y., Huenerfauth, M., Liu, J.: Recognizing American sign language gestures from within continuous videos. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2018)
Zeshan, U., Vasishta, M., Sethna, M.: Implementation of Indian sign language in educational setting. Asia Pacific Disab. Rehabil. J. 16(1), 16–40 (2005)
Zheng, L., Liang, B., Jiang, A.: Recent advances of deep learning for sign language recognition. In: 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Sydney, Australia, pp. 1–7 (2017). https://doi.org/10.1109/DICTA.2017.8227483
Acknowledgements
The authors would like to acknowledge Mrs. Sucheta S Saundankar, Principal, Padsad Karanbadhir Vidyalaya Nashik, and Mrs. P Jessy, Principal, Sheila Kothavala Institute for the Deaf, Bangalore, India, for allowing us to conduct the validation study at their respective schools. We are grateful to the Principal, Jessy Samuel and Speech Therapist, Ms. Suvasini Isaac for personally verifying the signs used in the validation stage for accuracy and consistency. We thank the anonymous reviewers whose insightful comments and suggestions have significantly improved this paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Mehta, N., Pai, S. & Singh, S. Automated 3D sign language caption generation for video. Univ Access Inf Soc 19, 725–738 (2020). https://doi.org/10.1007/s10209-019-00668-9
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10209-019-00668-9