Skip to main content

Advertisement

Log in

A novel method for head pose estimation based on the “Vitruvian Man”

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

In video-surveillance and ambient intelligence applications, head-pose estimation is an important and challenging task. Basically, the problem lies in assessing the pose of the head according to three reference angles, that indicate the head rotation with respect to three orthogonal axes, and are named roll, yaw, and pitch angles. The problem becomes particularly difficult if only 2D video-sequences or still images are available, thus information about the depth of the scene is missing. This makes the computation of the pitch angle very difficult. State-of-the-art methods usually add the information on the pitch angle separately, and this makes them strongly dependent on the hardware used and the scene under surveillance. Moreover, some of them require large training sets with head poses data. Finally, the extraction of several features from the detected face is often necessary. Since head-pose estimation is only a (small) part of a video-surveillance system as a whole, it is necessary to find novel approaches which make the head-pose estimation as simple as possible, in order to allow their use in real-time. In this paper, a novel method for automatic head-pose estimation is presented. This is based on a geometrical model relying on the exploitation of the Vitruvian man’s proportions and the related “Golden Ratio”. Our approach reduces the number of features extracted, avoiding the need for a training set as well as information on the hardware used or the scene under control. Simple ratios among eyes and nose positions, according to the assumed “Golden Ratio”, are used to compute, in particular, the pitch angle. Proposed method performs competitively with respect to state-of-the-art approaches, without requiring their working constraints and assumptions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Notes

  1. We did not report other methods because the strong performance difference is evident from [20].

  2. Greek letter ϕ recalls the initials of the sculptor Phidias, who used the “Golden Ratio” to create the Parthenon sculptures.

  3. An example of real-time use of our system can be found on the web at http://prag.diee.unica.it/amilab/en/node/60

References

  1. Asteriadis S, Soufleros D, Karpouzis K, Kollias S (2009) A natural head pose and eye gaze dataset. In: International conference on multimodal interfaces (ICMI 2009). doi:10.1145/1655260.1655261

  2. Asteriadis S, Karpouzis K, Kollias S (2010) Head pose estimation with one camera, in uncalibrated environments. Int.l Work. on Eye Gaze in Intelligent Human Machine Interactio, pp 55–62. doi:10.1145/2002333.2002343

  3. Baker BW, Woods MG (2001) The role of the divine proportion in the esthetic improvement of patients undergoing combined orthodontic/orthognathic surgical treatment. Int J Adult Orthodon Orthognath Surg 16(2):108–120

    Google Scholar 

  4. Cascia ML, Sclaroff S, Athitsos V (2000) Fast, reliable head tracking under varying illumination: an approach based on registration of texture-mapped 3D models. IEEE Trans Pattern Anal Mach Intell 22:322–336

    Article  Google Scholar 

  5. Computer Vision Centre , Universitat Autònoma de Barcelona (2006) Human expressive representations of motion and their evaluation in sequences. http://www.hermes-project.eu

  6. da Vinci L (1490) Vitruvian Man, Gallerie dell’Accademia. Venice, Italy, ca. 1490. http://brunelleschi.imss.fi.it/stampa_leonardo/images/uomo_vitruviano_accademia_v.jpg

  7. Gee AH, Cipolla R (1994) Determining the gaze of faces in images. Image Vis Comput 12(10):639–647

    Article  Google Scholar 

  8. Goldstein RE 1998. Esthetics in dentistry, B.C. Decker Inc

  9. Gong X, Li X-X, Feng L, Xia R (2012) A robust framework for face contour detection from clutter background. Int J Mach Learn Cybern 3(2):111–118

    Article  Google Scholar 

  10. Gourier J, Hall D (2004) Estimating face orientation from robust detection of salient facial features. In: Proc. of Pointing 2004, ICPR, international workshop on visual observation of deictic gestures, Cambridge, UK

  11. Gourier N, Maisonnasse J, Hall D, Crowley J (2006) Head pose estimation on low resolution images, multimodal technologies for perception of humans, international workshop classification of events activities and relationships, CLEAR 2006, LNCS R. Stiefelhagen and J. Garofolo, Eds., vol. 4122, pp. 270–280

  12. Habbema L (2004) Facial esthetics and patient selection. Clin Dermatol 22(1):14–7

    Article  Google Scholar 

  13. Hansen DW, Ji Q (2010) In the eye of the beholder: a survey of models for eyes and gaze. IEEE Trans Pattern Anal Mach Intell 32(3):478–500

    Article  Google Scholar 

  14. Horprasert T, Yacoob Y, Davis LS (1996) Computing 3-d head orientation from a monocular image sequence. In: Proc. of the 2nd international conference on automatic face and gesture recognition (FG ’96), pp 242–347. doi:10.1109/AFGR.1996.557271

  15. Hu Y, Chen L, Zhang H (2004) Estimating face pose by facial asymmetry and geometry. In: Proc. of sixth IEEE international conference on automatic face and gesture recognition, pp 651–656

  16. Lefèvre S, Odobez JM (2009) Structure and appearance features for robust 3D facial actions tracking. In: Proceedings of international conference on multimedia and expo, pp 298–301. doi:10.1109/ICME.2009.5202494

  17. Lefèvre S, Odobez JM (2010) View-based appearance model online learning for 3D deformable face tracking. In: Proceedings of international conference on computer vision theory and applications, pp 223–230

  18. Moeslund TB, Mortensen BK, Hansen DM (2006) Detecting head orientation in low resolution surveillance video, Technical report, CVMT-06-02 ISSN 1601-3646

  19. Morency L-P, Whitehill J, Movellan J (2008) Generalized adaptive view-based appearance model: integrated framework for monocular head pose estimation. In: IEEE international conference on automatic face and gesture recognition (FG), pp 1–8. doi:10.1109/AFGR.2008.4813429

  20. Murphy-Chutorian E, Trivedi M (2009) Head pose estimation in computer vision: a survey. IEEE Trans Pattern Anal Mach Intell 31(4):607–626

    Article  Google Scholar 

  21. Park S, Aggarwal J (2000) Head segmentation and head orientation in 3D space for pose estimation of multiple people. In: Proceedings of IEEE S symposium on image analysis and interpretation (SSIAI), pp 192–196

  22. Perseo G (2004) The “Beauty” of Homo Sapiens Sapiens: standard canons, ethnical, geometrical and morphological facial biotypes, Virtual J Orthod 10–5 (4). http://www.vjo.it/054/enface.pdf

  23. Poon B, Ashraful Amin M, Yan H (2011) Performance evaluation and comparison of PCA based human face recognition methods for distorted images. Int J Mach Learn Cybern 2(4):245–259

    Article  Google Scholar 

  24. Ricketts RM (1982) Divine proportion in facial esthetics. Clin Plast Surg 9(4):401–22

    Google Scholar 

  25. Sankaran P, Gundimada S, Tompkins RC, Asari VK (2005) Pose angle determination by face, eyes and nose localization. In: Proceedings of the 2005 IEEE conference on computer vision and pattern recognition (CVPR’05), p 161. doi:10.1109/CVPR.2005.582

  26. Stiefelhagen R (2004) Estimating head pose with neural networks—results on the Pointing04 - ICPR workshop evaluation data. In: Proceedings of Pointing 2004 Workshop: ICPR, Visual Observation of Deictic Gestures, Cambridge, UK

  27. Vatahska T (2007) Feature-based head pose estimation from images. In: Proceedings of 7th IEEE-RAS international conference on humanoid robots, 2007, pp 330–335

  28. Viola P, Jones M (2004) Robust real-time face detection. Int J Comput Vis 57(2):137–154

    Article  Google Scholar 

  29. Wang J, Sung E (2007) EM enhancement of 3D head pose estimated by point at infinity. Image Vis Comput 25(12):1864–1874

    Article  Google Scholar 

  30. Xiao J, Moriyama T, Kanade T, Cohn J (2003) Robust full-motion recovery of head by dynamic templates and re-registration techniques. Int J Imaging Syst Technol 13(1):8594

    Article  Google Scholar 

  31. Xu X, Liu W, Venkatesh S (2012) An innovative face image enhancement based on principle component analysis. Int J Mach Learn Cybern 3(4):259–0267

    Article  Google Scholar 

Download references

Acknowledgments

This work has been partially supported by Regione Autonoma della Sardegna (Legge Regionale 7 Agosto 2007, N. 7: “Promozione della ricerca scientifica e dell’ innovazione tecnologica in Sardegna”). Project ID CRP2-442.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gian Luca Marcialis.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Marcialis, G.L., Roli, F. & Fadda, G. A novel method for head pose estimation based on the “Vitruvian Man”. Int. J. Mach. Learn. & Cyber. 5, 111–124 (2014). https://doi.org/10.1007/s13042-013-0188-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-013-0188-y

Keywords

Navigation