Abstract
In this paper, we present a three-stage approach, which creates realistic facial animations by tracking expressions of a human face in 2D and transferring them to a human-like 3D model in real-time. Our calibration-free method, which is based on an average human face, does not require training. The tracking is performed using a single camera to enable several practical applications, for example, using tablets and mobile devices, and the expressions are transferred with a joint-based system to improve the quality and persuasiveness of animations. In the first step of the method, a joint-based facial rig providing mobility to pseudo-muscles is attached to the 3D model. The second stage covers the tracking of 2D positions of the facial landmarks from a single camera view and transfer of 3D relative movement data to move the respective joints on the model. The last step includes the recording of animation using a partially automated key-framing technique. Experiments on the extended Cohn-Kanade dataset using peak frames in frontal-view videos have shown that the presented method produces visually satisfying facial animations.
Similar content being viewed by others
Notes
Video: (https://vimeo.com/153692017)
References
Al-Nuaimi T (2006) Face recognition and computer graphics for modelling expressive faces in 3D. PhD thesis, Massachusetts Institute of Technology
Autodesk (2015) Makehuman. https://www.autodesk.com/education/free-software/maya
Black MJ, Yacoob Y (1995) Tracking and recognizing rigid and non-rigid facial motions using local parametric models of image motion. In: Proceedings of IEEE international conference on computer vision, pp 374–381
Cao C, Hou Q, Zhou K (2014) Displaced dynamic expression regression for real-time facial tracking and animation. ACM Trans Graph 33(4):1–10
Cao C, Wu H, Weng Y, Shao T, Zhou K (2016) Real-time facial animation with image-based dynamic avatars. ACM Trans Graph 35(4):1–12
Chai J, Xiao J, Hodgins J (2003) Vision-based control of 3d facial animation. In: Proceedings of the 2003 ACM SIGGRAPH / Eurographics symposium on computer animation, Eurographics Association, Aire-la-Ville, Switzerland, pp 193-206
Cohen M M, Massaro D W (1990) Synthesis of visible speech. Behav Res Methods Instrum Comput 22(2):260–263
Cohen M M, Massaro D W, et al. (1993) Modeling coarticulation in synthetic visual speech. In: Thalmann NM, Thalmann D (eds) Models and techniques in computer animation. Springer Japan, Tokyo, pp 139–156
Cootes T F, Edwards G J, Taylor C J (2001) Active appearance models. IEEE Trans Pattern Anal Mach Intell 23(6):681–685
Covell M (1996) Eigen-points: Control-point location using principal component analyses. In: Proceedings of the 2nd international conference on automatic face and gesture recognition, pp 122–127
DeCarlo D, Metaxas D (2000) Optical flow constraints on deformable models with applications to face tracking. Int J Comput Vis 38(2):99–127
Deng Z, Neumann U (2008) Data-driven 3D facial animation. Springer, London
Ekmen B, Ekenel HK (2016) Real time animated facial expression transfer. In: 2016 IEEE 24th signal processing and communication application conference (SIU), pp 1193–1196
Essa I, Basu S, Darrell T, Pentland A (1996) Modeling, tracking and interactive animation of faces and heads//using input from video. In: Computer animation ’96. Proceedings, pp 68–79
Faceshift (2017) Faceshift. http://faceshift.com/studio/2015.2/index.html
Gambaretto E, Piña C (2014) Real-time animation of cartoon character faces. In: ACM SIGGRAPH 2014 Computer Animation Festival, ACM, New York, NY, USA, SIGGRAPH ’14, p 1
Gokturk S B, Bouguet J Y, Grzeszczuk R (2001) A data-driven model for monocular face tracking. In: Proceedings 8th IEEE international conference on computer vision. ICCV 2001, vol 2, pp 701–708
Ichim A E, Bouaziz S, Pauly M (2015) Dynamic 3D avatar creation from hand-held video input. ACM Trans Graph 34(4):1–14
Kanade T, Cohn JF, Tian Y (2000) Comprehensive database for facial expression analysis. In: Proceedings 4th IEEE international conference on automatic face and gesture recognition (Cat. No. PR00580), pp 46–53
Lanitis A, Taylor C J, Cootes T F (1997) Automatic interpretation and coding of face images using flexible models. IEEE Trans Pattern Anal Mach Intell 19 (7):743–756
Laursen M H, Pedersen K S (2012) Partially automated system for synthesising human facial expressions in interactive media, Master’s thesis, Aalborg University
Lewis J P, Parke F I (1986) Automated lip-synch and speech synthesis for character animation. SIGCHI Bull 17(SI):143–147
Lewis JP, Pighin F (2006) Retargeting: Algorithms for performance-driven animation. In: ACM SIGGRAPH 2006 Courses, ACM, New York, NY, USA, SIGGRAPH ’06
Li L, Liu Y, Zhang H (2012) A survey of computer facial animation techniques. In: 2012 International conference on computer science and electronics engineering, vol 3, pp 434–438
Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Matthews I (2010) The extended cohn-kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society conference on computer vision and pattern recognition - workshops, pp 94–101
MakeHuman (2015) Makehuman. http://www.makehuman.org
Noh J, Neumann U (1998) A survey of facial modeling and animation techniques. Tech. rep
Parke FI (1974) A parametric model for human faces. Tech. rep
Parke FI, Waters K (2008) Computer facial animation. CRC Press, New York
Pelachaud C, Badler N I, Steedman M (1996) Generating facial expressions for speech. Cogn Sci 20(1):1–46
Pighin F, Szeliski R, Salesin D H (1999) Resynthesizing facial animation through 3d model-based tracking. In: Proceedings of the 7th IEEE international conference on computer vision, vol 1, pp 143–150
Platt S M, Badler N I (1981) Animating facial expressions. SIGGRAPH Comput Graph 15(3):245–252
Ruhland K, Prasad M, McDonnell R (2017) Data-driven approach to synthesizing facial animation using motion capture. IEEE Comput Graph Appl 37 (4):30–41
Singular Inversions (2015) Facegen modeller. https://facegen.com
Sreeja P S, Mahalakshmi G S (2017) Emotion models: a review. Int J Control Theory Appl 10:651–657
Terzopoulos D, Waters K (1993) Analysis and synthesis of facial image sequences using physical and anatomical models. IEEE Trans Pattern Anal Mach Intell 15 (6):569–579
Thies J, Zollhöfer M, Stamminger M, Theobalt C, Nießner M (2016) Face2Face: Real-time face capture and reenactment of RGB videos. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 2387–2395
Villagrasa S, Susín Sánchez A (2009) Face! 3D facial animation system based on facs. In: IV Iberoamerican symposium in computer graphics, pp 203–209
Waters K (1987) A muscle model for animation three-dimensional facial expression. SIGGRAPH Comput Graph 21(4):17–24
Williams L (1990) Performance-driven facial animation. SIGGRAPH Comput Graph 24(4):235–242
Xiong X, la Torre FD (2013) Supervised descent method and its applications to face alignment. In: 2013 IEEE conference on computer vision and pattern recognition, pp 532–539
Yehia H, Rubin P, Vatikiotis-Bateson E (1998) Quantitative association of vocal-tract and facial behavior. Speech Comm 26(1):23–43
Zhao H, Tai C L (2007) Subtle facial animation transfer from 2D videos to 3D faces with laplacian deformation. Proceedings of computer animation and social agents, Hasselt, Belgium
Acknowledgements
This work was supported by the TÜBİTAK project 113E067 and the EU Seventh Framework Programme Marie Curie FP7 integration project.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Ekmen, B., Ekenel, H.K. From 2D to 3D real-time expression transfer for facial animation. Multimed Tools Appl 78, 12519–12535 (2019). https://doi.org/10.1007/s11042-018-6785-8
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-018-6785-8