Abstract
This paper presents a novel algorithm for the extraction of the eye and mouth (facial features) fields from 2D gray level images. Eigenfeatures are derived from the eigenvalues and eigenvectors of the binary edge data set constructed from eye and mouth fields. Such eigenfeatures are ideal features for finely locating fields efficiently. The eigenfeatures are extracted from a set of the positive and negative training samples for facial features and are used to train a multilayer perceptron (MLP) whose output indicates the degree to which a particular image window contains the eyes or the mouth within itself. An ensemble network consisting of a multitude of independent MLPs was used to enhance the generalization performance of a single MLP. It was experimentally verified that the proposed algorithm is robust against facial size and even slight variations of the pose.
Similar content being viewed by others
References
Yankang Wang, H. Kuroda, M. Fujumura, and A. Nakamura, “Automatic extraction of eye and mouth fields from monochrome face image using fuzzy technique,” in Proceedings of the 1995 4th IEEE International Conference on Universal Personal Communications Record, 1995, pp. 778–782.
R. Pinto-Elias and J.H. Sossa-Azuela, “Automatic facial feature detection and location,” in Proceedings of the 14th International Conference on Pattern Recognition, 1998, vol. 2, pp. 1360–1364.
K. Fukui and O. Yamaguchi, “Facial feature point extraction method based on combination of shape extraction and pattern matching,” Systems and Computers in Japan, vol. 29, no. 6, pp. 49–58, 1998.
T. Kanade, “Picture processing by computer complex and recognition of human faces,” Dept. of Information Science, Kyoto University, Technical Report, 1973.
R. Brunelli and T. Poggio, “Face recognition: Features versus templates,” IEEE Transaction on PAMI, vol. 15, no. 10, pp. 1042–1052, 1993.
D.J. Beymer, “Face recognition under varying pose,” MIT A.I Memo No. 1461, 1993.
P. Juell and R. Marsh, “A hierarchical neural network for human face detection,” Pattern Recognition, vol. 29, no. 5, pp. 781–787, 1996.
Fenghao Mu, Haibo Li, and R. Forchheimer, “Automatic extraction of human facial features,” Signal Processing: Image Communication, vol. 8, pp. 309–326, 1996.
T. Shakunaga, K. Ogawa, and S. Oki, “Integration of eigentemplate and structure matching for automatic facial feature detection,” in Proceeding of the 2nd International Conference on Automatic Face and Gesture Recognition, 1998, pp. 94–99.
Ying-li Tian, T. Kanade, and J.F. Cohn, “Recognizing lower face action units for facial expression analysis,” in Proceeding of the 4th International Conference on Automatic Face and Gesture Recognition, 2000, pp. 484–490.
Ying-li Tian, T. Kanade, and J.F. Cohn, “Dual-state parametric eye tracking,” in Proceeding of the 4th International Conference on Automatic Face and Gesture Recognition, 2000, pp. 110–115.
R. Battiti and A.M. Colla, “Democracy in neural nets: Voting schemes for classification,” Neural Networks, vol. 7, no. 4, pp. 691–707, 1994.
D.F. McCoy and V. Devarajan, “Artificial immume systems and aerial image segmentation,” IEEE International Conference on Systems, Man, and Cybernetics, 1997.
M. Riedmiller and H. Braun, “Adirect adaptive method for faster backpropagation learning: The RPROP algorithm,” IEEE International Conference on Neural Networks, 1993, vol. 1, pp. 586–591.
E. Fiesler and R. Beale, Handbook of Neural Computation, Oxford University Press, 1997.
S. Romdhani, “ Face recognition using principal components analysis, ” MS Thesis, 1996, Available at http://www.elec.gla.ac.uk/(romdhani)/pca.htm.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Ryu, YS., Oh, SY. Automatic Extraction of Eye and Mouth Fields from a Face Image Using Eigenfeatures and Ensemble Networks. Applied Intelligence 17, 171–185 (2002). https://doi.org/10.1023/A:1016160814604
Issue Date:
DOI: https://doi.org/10.1023/A:1016160814604