Skip to main content

A Study Examining a Real-Time Sign Language-to-Text Interpretation System Using Crowdsourcing

  • Conference paper
  • First Online:
Computers Helping People with Special Needs (ICCHP 2020)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 12377))

Abstract

The current study examined how to use crowdsourcing to convert sign language-to-text. Generally in Japan, a sign language interpreter reads and vocalizes the sign language of the speaker, and caption typists generate captions from the vocalization. However, this method doubles labor costs and delays caption provision. Therefore, we developed a system that interprets sign language-to-caption text via crowdsourcing, with non-experts performing interpretations. While many individuals classified as deaf/hard-of-hearing (DHH) who can read sign language are suitable for this task, not all of them possess adequate typing skills. To address this, our system divides live sign language video into shorter segments, distributing them to workers. After the worker interprets and types the segments to text, the system generates captions through integration of these texts. Furthermore, we provide a user interface for playback speed control and one second rewinding in order to improve the ease with which tasks are completed. Our system can establish an environment that not only allows the interpretation of sign language-to-caption text, but also provides an opportunity for DHH individuals to assist those that are unable read sign language. We conducted a test using our prototype system for sign language-to-text interpretation. The mean time it took a worker to finish a task was 26 s for a 9 s segment. The combined total rate of missing text and collision between segments was 66%. Analysis of questionnaire responses found that workers assigned fewer tasks considered the tasks more enjoyable.

This study was partially supported by JSPS KAKENHI Grant Number JP15K01056 and JP19K02996.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 54.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 69.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Preminger, J.E., Levitt, J.E.: Computer-Assisted Remote Transcription (CART): a tool to aid people who are deaf or hard of hearing in the workplace. Volta Rev. 99(4), 219–230 (1997)

    Google Scholar 

  2. Elliot, L.B., Stinson, M.S., McKee, B.G., Everhart, V.S., Francis, P.J.: College students’ perceptions of the c-print speech-to-text transcription system. J. Deaf Stud. Deaf Educ. 6(4), 285–298 (2001)

    Article  Google Scholar 

  3. Kurita, S.: Development of CART software “IPtalk” and captioning services via personal computers: Communication support for the hearing-impaired. J. Inf. Process. Manag. 59(6), 366–376 (2016). https://doi.org/10.1241/johokanri.59.366

    Article  Google Scholar 

  4. Wakatsuki, D., Kato, N., Shionome, T., Kawano, S., Nishioka, T., Naito, I.: Development of web-based remote speech-to-text interpretation system captionline. JACIII 21(2), 310–320 (2017). https://doi.org/10.20965/jaciii.2017.p0310

    Article  Google Scholar 

  5. Lasecki, W.S., et al.: Real-time captioning by groups of non-experts. In: Proceedings of the ACM Symposium on User Interface Software and Technology (UIST 2012), pp. 23–34 (2012)

    Google Scholar 

  6. Lasecki, W.S., Miller, C.D., Bigham, J.P.: Warping time for more effective real-time crowdsourcing. In: Proceedings of the International ACM Conference on Human Factors in Computing Systems (CHI 2013), pp. 2033–2036 (2013)

    Google Scholar 

  7. Shiraishi, Y., Zhang, J., Wakatsuki, D., Kumai, K., Morishima, A.: Crowdsourced real-time captioning of sign language by deaf and hard-of-hearing people. Int. J. Pervasive Comput. Commun. 13(1), 2–25 (2017)

    Article  Google Scholar 

  8. Kumai, K., Zhang, J., Shiraishi, Y., Wakatsuki, D., Kitagawa, H., Morishima, A.: Group rotation management in real-time crowdsourcing. In: Proceedings of 19th International Conference on Information Integration and Web-based Applications & Services (iiWAS 2017), pp. 23–31 (2017)

    Google Scholar 

  9. Kumai, K., et al.: Skill-and-stress-aware assignment of crowd-worker groups to task streams. In: The sixth AAAI Conference on Human Computation and Crowdsourcing (HCOMP2018), pp. 88–97 (2018)

    Google Scholar 

  10. Shionome, T., et al.: Complement of incomplete task results for real-time crowdsourcing interpretation. In: Proceedings of 21th International Conference on Asian Language Processing (IALP 2017), pp. 359–362 (2017)

    Google Scholar 

  11. Hashimoto, H., Matsubara, M., Shiraishi, Y., Wakatsuki, D., Zhang J., Morishima, A.: A task assignment method considering inclusiveness and activity degree. In: Proceedings of The Second IEEE Workshop on Human-in-the-loop Methods and Human Machine Collaboration in BigData (IEEE HMData2018), pp. 3498–3503 (2018)

    Google Scholar 

  12. YouTube Live. https://www.youtube.com/live?gl=JP&hl=ja. Accessed 4 June 2020

  13. Node.js. https://nodejs.org/. Accessed 4 June 2020

  14. YouTube Player API. https://developers.google.com/youtube/iframe_api_reference?hl=ja. Accessed 4 June 2020

Download references

Acknowledgements

We would like to thank Editage (www.editage.com) for English language editing.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kohei Tanaka .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tanaka, K., Wakatsuki, D., Minagawa, H. (2020). A Study Examining a Real-Time Sign Language-to-Text Interpretation System Using Crowdsourcing. In: Miesenberger, K., Manduchi, R., Covarrubias Rodriguez, M., Peňáz, P. (eds) Computers Helping People with Special Needs. ICCHP 2020. Lecture Notes in Computer Science(), vol 12377. Springer, Cham. https://doi.org/10.1007/978-3-030-58805-2_22

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-58805-2_22

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-58804-5

  • Online ISBN: 978-3-030-58805-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics