Skip to main content

Automatic User-Specific Avatar Parametrisation and Emotion Mapping

  • Conference paper
Advanced Concepts for Intelligent Vision Systems (ACIVS 2013)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 8192))

  • 3233 Accesses

Abstract

In this paper an approach for automatic user-specific 3D model generation and expression classification is proposed. User performance-driven avatar animation is recently in the focus of research due to the increasing amount of low-cost acquisition devices with integrated depth map computation. Thereby challenging is the user-specific emotion classification without a complex manual initial step. Correct classification and emotion intensity identification can only be done with known expression specific facial feature displacement which differs from user to user. The use of facial feature tracking on predefined 3D model expression animations is presented here as solution statement for automatic emotion classification and intensity calculation. Consequently with this approach partial occlusions of a presented emotion do not hamper expression identification based on the symmetrical structure of human faces. Thus, a markerless, automatic and easy to use performance-driven avatar animation approach is presented.

The original version of this chapter was revised: The copyright line was incorrect. This has been corrected. The Erratum to this chapter is available at DOI: 10.1007/978-3-319-02895-8_64

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Bassili, J.N.: Facial motion in the perception of faces and of emotional expression. Journal of Experimental Psychology: Human Perception and Performance 4(3), 373–379 (1978)

    Google Scholar 

  2. Bouguet, J.-Y.: Pyramidal implementation of the affine lucas kanade feature tracker description of the algorithm. Intel Corporation (2001)

    Google Scholar 

  3. Chai, J.-X., Xiao, J., Hodgins, J.: Vision-based control of 3d facial animation. In: Proceedings of the 2003 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, pp. 193–206. Eurographics Association (2003)

    Google Scholar 

  4. Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. Journal of Personality and Social Psychology 17(2), 124 (1971)

    Article  Google Scholar 

  5. Facegen modeller (April 2013), http://facegen.com/modeller.htm

  6. Fasel, B., Luettin, J.: Automatic facial expression analysis: a survey. Pattern Recognition 36(1), 259–275 (2003)

    Article  MATH  Google Scholar 

  7. Joshi, P., Tien, W.C., Desbrun, M., Pighin, F.: Learning controls for blend shape based realistic facial animation. In: ACM SIGGRAPH 2005 Courses, p. 8. ACM (2005)

    Google Scholar 

  8. Lewis, J., Anjyo, K.-I.: Direct manipulation blendshapes. IEEE Computer Graphics and Applications 30(4), 42–50 (2010)

    Article  Google Scholar 

  9. Lienhart, R., Kuranov, A., Pisarevsky, V.: Empirical analysis of detection cascades of boosted classifiers for rapid object detection. In: Michaelis, B., Krell, G. (eds.) DAGM 2003. LNCS, vol. 2781, pp. 297–304. Springer, Heidelberg (2003)

    Chapter  Google Scholar 

  10. Liu, X., Xia, S., Fan, Y., Wang, Z.: Exploring non-linear relationship of blendshape facial animation. In: Computer Graphics Forum., vol. 30, pp. 1655–1666. Wiley Online Library (2011)

    Google Scholar 

  11. Niese, R., Al-Hamadi, A., Michaelis, B.: A novel method for 3d face detection and normalization. Journal of Multimedia 2(5), 1–12 (2007)

    Article  Google Scholar 

  12. Pantic, M., Rothkrantz, L.J.M.: Automatic analysis of facial expressions: The state of the art. IEEE Transactions on Pattern Analysis and Machine Intelligence 22(12), 1424–1445 (2000)

    Article  Google Scholar 

  13. Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, vol. 1, pp. I–511. IEEE (2001)

    Google Scholar 

  14. Weise, T., Bouaziz, S., Li, H., Pauly, M.: Realtime performance-based facial animation. ACM Trans. Graph. 30(4), 77 (2011)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Behrens, S., Al-Hamadi, A., Niese, R., Redweik, E. (2013). Automatic User-Specific Avatar Parametrisation and Emotion Mapping. In: Blanc-Talon, J., Kasinski, A., Philips, W., Popescu, D., Scheunders, P. (eds) Advanced Concepts for Intelligent Vision Systems. ACIVS 2013. Lecture Notes in Computer Science, vol 8192. Springer, Cham. https://doi.org/10.1007/978-3-319-02895-8_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-02895-8_18

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-02894-1

  • Online ISBN: 978-3-319-02895-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics