Kernel Discriminant Embedding in face recognition

https://doi.org/10.1016/j.jvcir.2011.07.009Get rights and content

Abstract

In this paper, we present a novel and effective feature extraction technique for face recognition. The proposed technique incorporates a kernel trick with Graph Embedding and the Fisher’s criterion which we call it as Kernel Discriminant Embedding (KDE). The proposed technique projects the original face samples onto a low dimensional subspace such that the within-class face samples are minimized and the between-class face samples are maximized based on Fisher’s criterion. The implementation of kernel trick and Graph Embedding criterion on the proposed technique reveals the underlying structure of data. Our experimental results on face recognition using ORL, FRGC and FERET databases validate the effectiveness of KDE for face feature extraction.

Highlights

► Incorporation of kernel trick, Graph Embedding and Fisher criteria in KDE. ► Kernel trick reveals underlying nonlinear data structure. ► KDE learns data from neighbourhood geometrical structure. ► The intrinsic structures implicitly exhibit certain degree of discriminating power. ► Fisher’s criterion in KDE explicitly enhances the discriminating capability.

Introduction

Face recognition has received extensive attention from the public because of its wide applications in our daily life such as identity verification, human–computer interaction, surveillance, etc. The most popular technique in face recognition is the subspace based approach. This subspace approach stems from Turk and Pentland’s pioneering work which is known as Eigenface [1]. Essentially, Eigenface seeks an optimal linear transformation which aims to preserve the maximum variance of data. Fisherface (or Linear/Fisher Discriminant Analysis, L/FDA) is an improved version of Eigenface with incorporation of Fisher’s criterion. Fisherface method seeks a projection direction which maximizes the between-class scatter and minimizes the within-class scatter [2]. Typically, both techniques encode data using second order dependencies, i.e. pixel-wise scatter of the image data. Hence, they are insensitive to higher order dependencies which encode nonlinear relations among the pixels [3]. Therefore, Eigenface and Fisherface could be suboptimal for face recognition. Moreover, these techniques assume that data is inherently Gaussian which is usually not assured in practice.

Recent studies assume that high dimensional data is a set of geometrically associated points lying on or nearly on a low dimensional manifold [4], [5], [6]. Hence, the intrinsic geometrical structures of data may possess inherited discriminating power for classification. Some example algorithms, which reveal this intrinsic manifold structure of data, include the Laplacianface (or Locality Preserving Projection, LPP) [7] and the Neighbourhood Preserving Embedding (NPE) [8]. LPP preserves the neighbourhood structure of data set optimally based on a heat kernel nearest neighbour graph [7]. NPE takes into account the restriction that neighbouring points in the high dimensional space must remain within the same neighbourhood in the low dimension space and be located in a similar relative spatial situation (without changing the local structure of the nearest neighbours of each data point). Even though these methods are differentiated by their objective functions, they can be unified within Graph Embedding (GE) framework [9]. GE characterizes data points based on preserving properties of small neighbourhoods around the data points. The local neighbourhood relations illustrate the manifold embedding without any specific assumption on data distribution.

The underlying discriminating power of these algorithms may not always be assured since the real world data could be too complicated. In order to enhance the discriminating power of GE algorithms, a discriminant criterion has been explicitly integrated into the learning algorithms. For examples, Marginal Fisher Analysis (MFA) [9], Locality Sensitive Discriminant Analysis (LSDA) [10] and Neighbourhood Preserving Discriminant Embedding (NPDE) [11] incorporate the Fisher’s criterion (FC) for manifold learning. The Maximum Neighbourhood Margin Criterion (MNMC) as shown in [12] performs a manifold learning based on Maximum Margin Criterion (MMC). The promising discriminating capacity of these algorithms has been validated from the experimental results [9], [12]. However, these methods remained encoding pattern information based on second order dependencies; whereas those higher order dependencies in an image, such as the correlations among three or more pixels of an edge, which might capture pertinent information for recognition have been neglected [3]. Hence, a nonlinear mapping could be used to map the data in the input space to a higher dimensional feature space. The main purpose of mapping is to “unfold” the data manifold, so that those discriminative nonlinear structures can emerge under this new representation. The kernel trick allows this unfolding implicitly [3].

Section snippets

Motivation and contributions

In this paper, we present a novel technique for effective nonlinear discriminative feature extraction for face recognition. The proposed technique incorporates a kernel trick, a Graph Embedding (GE) and the Fisher’s criterion, called Kernel Discriminant Embedding (KDE). In KDE, the input data is first mapped into a higher dimensional feature space using the kernel trick. The kernel trick unfolds the data manifold, so that the underlying nonlinear features emerge. Then, the discriminant GE and

Kernel Discriminant Embedding

The objective of KDE is that the neighbourhood of points from the same class label is preserved, avoiding any data points from other classes to enter the neighbourhood. Hence, KDE utilizes the Fisher’s criterion in its feature extraction. KDE considers two distinct sets of reconstruction coefficients that can be computed from the face data of same and different identities: intra-class GE coefficients and inter-class GE coefficients for discrimination purpose. This proposed method assumes an

Justifications about KDE

For recognition purpose, the crucial part is to find a projection that possesses discriminating capability in such a way that it is able to minimize the intra-class scatter while maximizing the inter-class scatter. The proposed KDE accomplishes this principle to achieve a high discriminating power. To verify the idea empirically, we project the face images of two arbitrary picked subjects (denoted as class A and class B) in FRGC dataset (FRGC dataset will be described in later section) onto a

Experimental set-up

The effectiveness of the proposed KDE is evaluated for face recognition on (1) Olivetti Research Laboratory Database (ORL), (2) Face Recognition Grand Challenge Database (FRGC), and (3) FERET database.

The ORL database contains 400 images of 40 identities (10 images per person) [13]. The images in this database have minor variations in head poses and facial expressions (open or close eyes, smiling or non-smiling) as well as facial details (glasses or without glasses). Five sample face images

Concluding remarks

We have presented a Kernel Discriminant Embedding (KDE) for dimensionality reduction and feature extraction in face recognition. KDE is developed via conforming to three significant criteria which are helpful for discriminative feature extraction: kernel, Graph Embedding and the Fisher’s criteria. KDE is a linear projection method that allows directional projection of data for subspace analysis. By solving a linear eigenvalue problem in closed form, KDE automatically learns the neighbourhood

Acknowledgment

This work was supported by the Korea Science and Engineering Foundation (KOSEF) through the Biometrics Engineering Research Center (BERC) at Yonsei University (Grant No. R112002105080020 (2010)).

References (17)

  • Pang Ying Han et al.

    Neighbourhood preserving discriminant embedding in face recognition

    Elsevier J. Vis. Commun. Image Representation

    (2009)
  • M. Turk et al.

    Eigenfaces for recognition

    J. Cognitive Neurosci.

    (1991)
  • P.N. Belhumeur et al.

    Eigenfaces vs. Fisherfaces: recognition using class specific linear

    IEEE Trans. Pattern Anal. Machine Intell.

    (1997)
  • M.H. Yang, Kernel Eigenfaces vs. Kernel Fisherfaces: face recognition using kernel methods, in: Proceedings of the...
  • M. Belkin et al.

    Laplacian eigenmaps and spectral techniques for embedding and clustering

    Proc. Conf. Adv. Neural Inform. Process. System

    (2001)
  • S.T. Roweis et al.

    Nonlinear dimensionality reduction by locally linear embedding

    Science

    (2000)
  • J. Tenenbaum et al.

    A global geometric framework for nonlinear dimensionality reduction

    Science

    (2000)
  • X. He et al.

    Face recognition using laplacianfaces

    IEEE Trans. Pattern Anal. Machine Intell.

    (2005)
There are more references available in the full text version of this article.

Cited by (7)

  • Adjusting samples for obtaining better l<inf>2</inf>-norm minimization based sparse representation

    2016, Journal of Visual Communication and Image Representation
    Citation Excerpt :

    After determining the linear combination, SR then exploits it to classify the test sample, Huang et al. [15] utilized a multi-manifold metric learning based method to classify the test sample. Other classifying related method include feature extraction method, e.g., a kernel trick proposed by Han et al. [16], there are also methods considering the relationship between image resolution and face recognition [17]. The two-phase test sample SR previously proposed [10] is a typical example of l2-norm minimization based SR.

  • Robust gait recognition via discriminative set matching

    2013, Journal of Visual Communication and Image Representation
    Citation Excerpt :

    Hence, MCCA attempts to find such a new subspace in which the distances between the neighboring subspaces are maximized. Such neighborhood-based discriminative learning has already been proved to be effective for vector-based classification [22,23,28,31]. Compared with the conference paper [33], this journal version is extended as follows:

  • Manifold Adaptive Label Propagation for Face Clustering

    2015, IEEE Transactions on Cybernetics
  • Face recognition with manifold-based kernel discriminant analysis

    2012, Proceedings of the International Joint Conference on Neural Networks
View all citing articles on Scopus
View full text