Skip to main content

Correct Speech Visemes as a Root of Total Communication Method for Deaf People

  • Conference paper
Book cover Agent and Multi-Agent Systems. Technologies and Applications (KES-AMSTA 2012)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 7327))

  • 2070 Accesses

Abstract

Many deaf people are using lip reading as a main communication fiorm. A viseme is a representational unit used to classify speech sounds in the visual domain and describes the particular facial and oral positions and movements that occur alongside the voicing of phonemes. A design tool for creating correct speech visemes is designed. It’s composed of 5 modules; one module for creating phonemes, one module for creating 3D speech visemes, one module for facial expression and modul for synchronization between phonemes and visemes and lastly one module to generate speech triphones. We are testing the correctness of generated visemes on Slovak speech domains. The paper descriebes our developed tool.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hrúz, M., Krňoul, Z., Campr, P., Muller Ludek, M.S.: Tovards Automatic Annotation of Sing Language Dictionarz Corpora. In: TDS Pilsen, pp. 331–339 (2011)

    Google Scholar 

  2. Cai, Q., Gallup, D., Zhang, C., Zhang, Z.: 3D deformable face tracking with a commodity depth camera. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part III. LNCS, vol. 6313, pp. 229–242. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  3. Drahoš, P., Šperka, M.: Face Expressions Animation in e-Learning, http://lconf06.dei.uc.pt/pdfs/paper14.pdf

  4. Albrecht, J.: Haber, and H.-P. Seidel. Speech Synchronization for Physicsbased Facial Animation. In: Skala, V. (ed.) Proc. 10th Int. Conf. on Computer Graphics, Visualization and Computer Vision (WSCG 2002), pp. 9–16 (2002)

    Google Scholar 

  5. Ma, J., Cole, R., Pellom, B., Ward, W., Wise, B.: Accurate visible speech synthesis based on concatenating variable length motion capture data. IEEE

    Google Scholar 

  6. Wang, A., Emmi, M., Faloutsos, P.: Assembling an expressive facial animation system. In: Sandbox 2007: Proceedings of the 2007 ACM SIGGRAPH Symposium on Video Games, pp. 21–26. ACM Press (2007)

    Google Scholar 

  7. Ezzat, T., Geiger, G., Poggio, T.: Trainable videorealistic speech animation. In: SIGGRAPH 2002: Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, pp. 388–398. ACM Press (2002)

    Google Scholar 

  8. Cohen, M.M., Massaro, D.W.: Modeling coarticulation in synthetic visual speech. In: Magnenat Thalmann, N., Thalmann, D. (eds.) Models and Techniques in Computer Animation, pp. 139–156. Springer, Tokyo (1994)

    Google Scholar 

  9. Brand, M.: Voice puppetry. In: SIGGRAPH 1999: Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, pp. 21–28. ACM Press/Addison-Wesley Publishing Co. (1999)

    Google Scholar 

  10. Kim, I.-J., Ko., H.-S.: 3d lip-synch generation with data-faithful machine learning. Computer Graphics Forum. In: EUROGRAPHICS 2007, vol. 26(3) (2007)

    Google Scholar 

  11. Bregler, C., Covell, M., Slaney, M.: Video rewrite:driving visual speech with audio. In: Computer Graphics, Proc. SIGGRAPH 1997, pp. 67–74 (1997)

    Google Scholar 

  12. Meadow-Orlans, K.P., Mertens, D.M., Marilyn, S.-L.: Parents and their Deaf Children. Gallaudet University Press, Washington D.C. (2003)

    Google Scholar 

  13. Munoz-Baell, I.M., Ruiz, T.M.: Empowering the deaf; let the deaf be deaf. Journal Epidemiol Community Health 54(1), 40–44 (2000)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Pajorová, E., Hluchý, L. (2012). Correct Speech Visemes as a Root of Total Communication Method for Deaf People. In: Jezic, G., Kusek, M., Nguyen, NT., Howlett, R.J., Jain, L.C. (eds) Agent and Multi-Agent Systems. Technologies and Applications. KES-AMSTA 2012. Lecture Notes in Computer Science(), vol 7327. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-30947-2_43

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-30947-2_43

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-30946-5

  • Online ISBN: 978-3-642-30947-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics