Elsevier

Neurocomputing

Volume 188, 5 May 2016, Pages 160-166
Neurocomputing

A local spectral feature based face recognition approach for the one-sample-per-person problem

https://doi.org/10.1016/j.neucom.2014.10.111Get rights and content

Abstract

Face recognition for the one-sample-per-person problem has received increasing attention owing to its wide range of potential applications. However, since only one training image is available for each person, and the face images may have large appearance variations, how to achieve a high recognition accuracy is still a challenging work. In this paper, we propose a more accurate local spectral feature based face recognition approach for the one-sample-per-person problem. In the proposed algorithm, multi-resolution local spectral features are first extracted to represent the face images to enlarge the training set. A weaker classifier is then constructed based on the spectral features of each local region. Since a good diversity is observed for the outputs of the weaker classifiers, a strategy of classifier committee learning is adopted to combine the results obtained from different local spectral features. Moreover, inspired by the fact that the iterations are completely independent of each other, a scheme of multiple worker based parallel computing is designed to improve the loop speed by distributing iterations to the MATLAB workers simultaneously. Experimental results on the standard databases demonstrate the feasibility and effectiveness of the proposed method.

Introduction

In recent years, face recognition for the one-sample-per-person problem has received increasing attention owing to its wide range of potential applications, e.g., law enforcement, surveillance identification, forensic identification and access control, etc [1], [2]. It is well known that, for the traditional statistical learning methods, the recognition performances generally heavily rely on the number of training sample. Thus, the lack of enough training samples often results in poor generalization ability, or even failure, for these methods. Since only one training image is available for each person, and the face images may have large appearance variations in terms of expression, illumination, disguises, pose, and so on, the one-sample-per-person problem has become an extremely challenging work of face recognition [3], [4], [5], [6].

So far, many approaches have been developed to address the one-sample-per-person face recognition problem. Generally speaking, the common strategy to deal with the problem is to enlarge the training set by extracting the various discriminative features [7], generating the virtual samples [8], or constructing the discriminant model by means of the generic set [9]. A typical strategy of the discriminative feature extraction method is to employ the features from the local region [10]. A prominent advantage of using local representations is its fair robustness to variations in lighting, expression and occlusion. In order to capture the intra-class variations for each single sample, some researchers proposed to synthesize virtual samples by using the learned information [11], by means of the various transformation [12], or by rendering the recovered 3D face model [2], etc. Among these methods, the information in the frequency domain are frequently utilized to strengthen the recognition performance [13]. In [14], the frequency invariant features and the moment invariant features [14] are combined for face recognition with a single training sample. In [15], the one-sample-per-person problem is addressed via a fusion of the directionality of edges and the intensity facial features. Moreover, different recognition approaches, such as sparse representation [16], and linear regression [17], [18], are proposed for the one-sample-per-person problem.

Recently, we presented a very effective recognition method for the one-sample-per-person problem, named MR_2DLDA [19]. In the proposed method, multi-resolution spectral feature images are constructed to represent the face images, which greatly enlarge the training set. 2DLDA (two-dimensional linear discriminant analysis) is then applied on the spectral representations. Experimental results on multiple databases demonstrated the superiority of the spectral feature representation.

Inspired by the robustness of the local feature, and the superiority of the spectral feature representation, in this paper, we propose a more accurate local spectral feature based face recognition approach for the one-sample-per-person problem. One issue with the one-sample-per-person problem is that the number of training samples available is too few. In the proposed method, multi-resolution spectral features are first extracted and used as the representations of training face images by means of a method similar to [20]. Further, the spectral feature images are divided into patches with the same sizes. Then, the patches of the same position are collected, and used as the training samples of one weak classifier. Thereby, the size of the training set is greatly enlarged via the local spectral representation.

As we do not know exactly which patches, orientations and scales are robust for all testing images, an alternative approach is to use all of local spectral features in the decision-making process. In our method, each patch with a certain orientation and scale will form one weak classifier. In order to determine the classes of the testing images, a strategy of classifier committee learning (CCL) is designed further to combine the results obtained from different local spectral feature images. With the strategy of CCL, on the one hand, most of the correct categorizations can be retained. On the other hand, it is not necessary for us to choose the optimal patches and filters, which is a very difficult task for the one-sample-per-person problem. Using the above strategies, the negative effects caused by those unfavorable factors, such as variations of illumination and facial expression, can be alleviated greatly in face recognition. It should be pointed out that, for the MR_2DLDA method [19], the diversity of weak classifiers is mainly due to the scale variation of spectral features. Nevertheless, for the proposed method, besides the scale variation, the local features of each scale can also increase the diversity of weak classifiers. As a result, the final recognition results obtained by the CCL are significantly improved.

Because local spectral features are used in the proposed method, the iteration computation number of the block distances becomes increasingly large. Inspired by the fact that the iterations are completely independent each other, a scheme of multiple worker based parallel computing is designed to improve the loop speed by distributing iterations to the MATLAB workers simultaneously.

The contributions of the proposed algorithm can be summarized as two aspects. A multi-resolution local spectral feature representation is proposed to enlarge the training set, and increase the diversity of the weaker classifiers at the CCL stage. Moreover, a scheme of multiple worker based parallel computing is designed to decrease the training times. Experimental results on some standard databases demonstrate the feasibility and effectiveness of the proposed method.

The remainder of the paper is organized as follows. In Section 2, we present our proposed algorithm. Experimental results and related discussions are given in Section 3, and concluding remarks are presented in Section 4.

Section snippets

Local spectral feature image representation

Assume that there are L training images Ii(i=1,,L), and that each belongs to one subject. The training image is first pre-filtered to reduce the effect of illumination. The Fourier transform of the prefiltered image is then filtered by a set of Gabor filters with ns scales and no orientations [20]. Further, the corresponding amplitudes are computed as the spectral feature images. Thus, for the given Nf (i.e. ns×no) filters, Nf spectral feature images can be obtained for each training sample.

Databases and experimental set-up

The performance of our proposed method is evaluated on five standard face image databases: Extended Yale Face database B, PIE database, FERET face database, AR database, and LFWA database [22], [23], which are relatively large and widely used databases for face recognition.

The Extended Yale Face database B has 38 individuals and around 64 near frontal images under different illuminations per individual. The PIE database contains images of 68 individuals [24]. In the FERET face database, there

Conclusion

In this paper, we propose a more accurate local spectral feature based face recognition algorithm for the one-sample-image-per-person problem. The obtained labels indicated that the weaker classifiers have a good diversity. In addition, the recognition rates demonstrated that our proposed approach is more accurate than some recently reported methods. Moreover, the computation times indicate that the parallel computing can effectively decrease the training times.

Acknowledgments

The work was supported by grants from National Natural Science Foundation of China (Nos. 61370109, 61373098), a grant from Natural Science Foundation of Anhui Province (No. 1308085MF85), and 2013 Zhan-Li Sun׳s Technology Foundation for Selected Overseas Chinese Scholars from Department of Human Resources and Social Security of Anhui Province (Project name: Research on structure from motion and its application on 3D face reconstruction).

Zhan-Li Sun received the Ph.D. degree from the University of Science and Technology of China, in 2005.

Since 2006, he has worked with The Hong Kong Polytechnic University, Nanyang Technological University, and National University of Singapore. He is currently a Professor with School of Electrical Engineering and Automation, Anhui University, China. His research interests include machine learning, and image and signal processing.

References (27)

Cited by (0)

Zhan-Li Sun received the Ph.D. degree from the University of Science and Technology of China, in 2005.

Since 2006, he has worked with The Hong Kong Polytechnic University, Nanyang Technological University, and National University of Singapore. He is currently a Professor with School of Electrical Engineering and Automation, Anhui University, China. His research interests include machine learning, and image and signal processing.

Li Shang received the B.Sc. degree and M.Sc. degree in Xi׳an Mine University in June 1996 and June 1999, respectively. And in June 2006, she received the Doctor׳s degree in Pattern Recognition & Intelligent System in University of Science & Technology of China (USTC), Hefei, China. From July 1999 to July 2006, she worked at USTC, and applied herself to teaching. Now, she works at the Department of Communication Technology, Electronic Information Engineering College, Suzhou Vocational University. At present, her research interests include Image processing, Artificial Neural Networks and Intelligent Computing.

View full text