Skip to main content

Motion Dynamic Analysis of the Basic Facial Expressions

  • Conference paper
  • 2723 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 8445))

Abstract

In interaction systems, communication between user and the computer may be performed using a graphical display of human representation called avatar. This paper is focussed on the problem of facial motion analysis for human-like animation. Using similarities in motion data four criteria for characteristic points grouping (facial regions, movement directions, angles and distances) have been proposed. In order to estimate the number of clusters for selected facial expressions a dedicated algorithm has been applied. Based on the results of subjective assessment the most satisfying configuration of criteria, in terms of number of clusters and accuracy of emotions recognition, was a group of distance, region and angle between facial markers. In the result, the obtained groups may be used to simplify the number of control parameters necessary to synthesise facial expressions in virtual human systems. The final structure of the characteristic points can diminish overall computational resources usage by decreasing the number of points that need to be recalculated between animation phases. This is due to the fact, that the movement similarities were exploited to make the groups with the same properties be controlled by dominant markers.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Thomaz, A.L., Breazeal, C.: Understanding Human Teaching Behavior to Build More Effective Robot Learners. Artificial Intelligence Journal 172 (2008)

    Google Scholar 

  2. Ekman, P., Friesen, W.V.: Unmasking the Face. A guide to recognizing emotions from facial expressions. Malor Books (2003)

    Google Scholar 

  3. Ekman, P.: Emotions Revealed: Recognizing Faces and Feelings to Improve Communication and Emotional Life. Owl Books (2007)

    Google Scholar 

  4. Mana, N., Pianesi, F.: Modelling of Emotional Facial Expressions during Speech in Synthetic Talking Heads using a Hybrid Approach. Auditory-Visual Speech Processing (2007)

    Google Scholar 

  5. Wik, P., Hjalmarsson, A.: Embodied conversational agents in Computer Assisted Language Learning. Speech Communication 51(10) (2009)

    Google Scholar 

  6. Breazeal, C.: Designing Sociable Robots. MIT Press (2002)

    Google Scholar 

  7. Zhao, H., Wang, Z., Men, J.: Facial Complex Expression Recognition Based on Fuzzy Kernel Clustering and Support Vector Machines. In: Third International Conference on Natural Computation (2007)

    Google Scholar 

  8. Zhang, S.-C., Fang, B., Liang, Y.-Z., Jing, W., Wu, L.: A Face Clustering Method Based on Facial Shape Information. In: Proceedings of the 2011 International Conference on Wavelet Analysis and Pattern Recognition, pp. 44–49 (2011)

    Google Scholar 

  9. Yu, X., Chen, X., Gao, M.: Motion Detection in Dynamic Scenes Based on Fuzzy C-Means Clustering. In: International Conference on Communication Systems and Network Technologies, pp. 306–310 (2012)

    Google Scholar 

  10. Lucero, J.C., Maciel, S.T., Johns, D.A., Munhall, K.G.: Empirical modeling of human face kinematics during speech using motion clustering. Journal of the Acoustical Society of America 118(1), 405–409 (2005)

    Article  Google Scholar 

  11. Ekman, P., Friesen, W.V., Hager, J.C.: Facial Action Coding System. The Manual. Research Nexus division of Network Information Research Corporation (2002)

    Google Scholar 

  12. Kocoń, M., Emirsajłow, Z.: Face Emotional States Mapping Based on the Rigid Bone Model. Journal of Applied Computer Science 19(2), 47–60 (2011)

    Google Scholar 

  13. Kocoń, M.: Idle Motion Synthesis of Human Head and Face in Virtual Reality Environment. In: Ma, M., Oliveira, M.F., Petersen, S., Hauge, J.B. (eds.) SGDA 2013. LNCS, vol. 8101, pp. 299–306. Springer, Heidelberg (2013)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Kocoń, M. (2014). Motion Dynamic Analysis of the Basic Facial Expressions. In: Likas, A., Blekas, K., Kalles, D. (eds) Artificial Intelligence: Methods and Applications. SETN 2014. Lecture Notes in Computer Science(), vol 8445. Springer, Cham. https://doi.org/10.1007/978-3-319-07064-3_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-07064-3_11

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-07063-6

  • Online ISBN: 978-3-319-07064-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics