Elsevier

Neurocomputing

Volume 173, Part 3, 15 January 2016, Pages 541-551
Neurocomputing

Discriminative low-rank dictionary learning for face recognition

https://doi.org/10.1016/j.neucom.2015.07.031Get rights and content

Abstract

Based on Low-Rank Representation (LRR), this paper presents a novel dictionary learning method to learn a discriminative dictionary which is more suitable for face recognition. Specifically, in order to make the dictionary more discriminating, we introduce an ideal regularization term with label information of training data to obtain low-rank coefficients. In the dictionary learning process, by optimizing the within-class reconstruction error and minimizing of between-class sub dictionaries, the learned dictionary has good representation ability for the training samples. In addition, we also suggest each sub dictionary is low-rank, which can violate with noise contained in training samples and make the dictionary more pure and compact. The learned dictionary and structured discriminative low-rank representation then will be used for classification. The proposed Discriminative Low-Rank Dictionary Learning (DLR_DL) method is evaluated on public face databases in comparison with previous dictionary learning under the same learning conditions. The experimental results demonstrate the effectiveness and robustness of our approach.

Introduction

In recent years, dictionary learning has been successfully applied in the field of computer vision. Many dictionary learning methods have been proposed and have achieved state-of-the-art performances for image processing [1], [2], [3] and classification [4], [5], [6], [7], [8], [9], [10], [11]. The dictionary learning is usually used for sparse representation or approximation of signals, and it aims to learn a dictionary from training samples in which only a few atoms can be linearly combined to well approximate given test samples.

The most basic way to build a dictionary is directly to use the training samples. Wright et al. [12] introduced sparse representation based classification (SRC) algorithm, where all the training samples are used to form a dictionary, and a test image is classified by finding its sparse representation with respect to this dictionary. Many methods have been proposed to improve the performance of SRC [13], [14], [15]. However, in these methods if the number of training images is very large, this will result in a big size dictionary which requires much memory and high computational cost. In addition, using original training samples as a dictionary could not exploit the discriminating information hidden in the training samples.

Several methods have been proposed to learn a small-sized dictionary from the training samples. KSVD [2] is proposed to efficiently learn an over-complete dictionary by updating dictionary atoms and sparse representations iteratively. This method has been applied in image compression and denoising. KSVD only considers the representational power of the learned dictionary but ignores its capability dicrimination. Marial et al. [4] proposed the dictionary learning model by adding a discriminative reconstruction constraint to gain the discriminative ability. Marial et al. [16] prpoposed an online dictionary learning algorithm for digit recognition and texture classification. By incorporating the classification error into the objective function, Zhang and Li [8] proposed discriminative KSVD (D-KSVD) algorithm for face recognition. Pham and Venkatesh [6] obtained a discriminative dictionary for object categorization and utilized a linear predictive classier to realize face recognition. The above methods learn a common dictionary shared by all classes as a classifier of coefficients for classification. However, the learned dictionary shared by all classes can easily lose the corresponding information between dictionary atoms and the class label, and hence they are not really effective while using the reconstruction error associated with each class for the classification. Some methods use the reconstruction error of each class as the discriminative information for classification are proposed. Yang et al. [11] learned a dictionary of each class for face recognition and achieved better results than SRC. Jiang et al. [17] add a label consistence term on K-SVD algorithm (LC-KSVD) to make the dictionary more discriminative for sparse coding. To reduce the computational complexity, Lee et al. [18] and Wang et al. [19] emphasized specific discriminative criteria to learn an over-complete dictionary. Ramirez et al. [9] learned the dictionaries with structure incoherence and shared features. Yang et al. [20] proposed a Fisher Discriminative Dictionary Learning (FDDL) for sparse representation. Studer and Baraniuk [21] investigated dictionary learning from sparsely corrupted signals. These above methods can learn concurrently the optimal dictionary and improve the discriminative capability of coding vectors. However, in the case of training samples corrupted with large noise, the dictionary atoms will also get corrupted.

In recent years, based on low-rank matrix recovery and completion [22], [23], [24], Wright et al. [25] proposed the Robust Principal Component Analysis (RPCA) to recover the underlying low-rank structure in the data. It has been successfully applied to applications including salient object detection [26], segmentation and grouping [27], [28], background subtraction [29], and tracking [30]. In image classification, Chen et al. [31] use low rank matrix recovery to remove noise from the training sample class by class for face recognition. Furthermore, Liu et al. [32], [33] established a low-rank representation (LRR) as an efficient way to perform noise correction and subspace segmentation simultaneously. Based on low-rank representation, a number of dictionary learning methods for image classification have been proposed. Ma et al. [34] presented a discriminative low-rank dictionary for sparse representation (DLRD_SR). By integrating rank minimization into sparse representation for dictionary learning, this method achieved impressive face recognition results especially when corruption existed. Zhang et al. [35] proposed a structure low-rank representation for image classification, by adding a regularization term to the objective function. Recently, Li et al. [36] proposed discriminative dictionary learning with low-rank regularization (D2L2R2) for image classification that can handle training samples corrupted with large noise. D2L2R2 combines the Fisher discrimination function with low-rank on the sub dictionary to make the learned dictionary more discerning and pure.

In this paper, we propose a Discriminative Low-Rank Dictionary Learning (DLR_DL) method for face recognition. To improve the discriminative power of the dictionary, we use an ideal regularization term with label information of training samples, which is incorporated into the objective function as a discrimination regularization term. When the training samples are corrupted with large noise, the dictionary atoms will get corrupted. Therefore, to reduce the effect of noise contained in training samples, we suggest each sub dictionary is low-rank, which can violate with noise contained in training samples and make the dictionary more pure and compact. In the dictionary learning process, we make each class specific in the whole structured discriminative dictionary which has good representation ability for the training samples from the associated class but poor for other class. The low-rank sparse coefficients and discriminate low-rank dictionary learned then will be used for classification.

Unlike DLRD_SR algorithm proposed in Ref. [34], the low-rank regularization can cause losses of certain information, however, our approach can still perform well by combining label information with low-rank representation. Different from the recently proposed algorithm in Ref. [35], which improves the representation power of the dictionary while ignoring its discriminative capacity, our method has superior performance by optimizing the reconstruction error of each class in the dictionary learning process. Compared with the SRC and DLRD_SR, DLR_DL not only get higher classification accuracy, but also learns a smaller dictionary. Experimental results show that our approach has competitive performance in pattern recognition tasks.

The remainder of this paper is organized as follows: Section 2 gives a brief review of some related work. Section 3 introduces a Discriminative Low-Rank Dictionary Learning approach. Section 4 presents the optimization of the proposed model. Section 5 describes the classification scheme. Experimental results are given in Section 6. Finally, Section 7 concludes this paper.

Section snippets

Low rank representation

Wright et al. [25] recently established the RPCA method aims to recover a low-rank matrix from corrupted input data. Suppose a matrix X can be decomposed into two matrices A and E, X=A+E, where A is low-rank matrix and E is a sparse matrix. Low-rank matrix recovery aims finding A and E from X. It can be viewed as an optimization problem: decomposing the input X into A+EminA,Erank(A)+λE0,subjecttoX=A+Ewhere λ>0 is a parameter that controls the weight of the noise matrix E. However, the rank

Discriminative low-rank dictionary learning

To improve the performance of dictionary learning algorithm when large noise exists in the training samples, we propose a low-rank discriminative dictionary learning scheme. In our proposed model, we desire that the structured dictionary D should have powerful discriminative and reconstructive capability of samples X. Given a set of training data vectorsX=[X1,X2,...,Xc]d×N, where Xi is the samples from class i, d is the feature dimension, and N is the number of total training samples. X may

Optimization of DLR_DL

The DLR_DL objective function in Eq. (8) can be divided into two sub-problems: updating Z by fixing D; and updating D by fixing Z. The procedures are iteratively implemented for the low-rank coefficients Z and the discriminative dictionary D. In the first sub-problem, we update Zi, (i=1, 2,…,c) one by one by fixing D and Zj (ji), then we get the coefficient matrix Z=[Z1;Z2;...;Zc]. In the second sub-problem, we update Di by fixing Zj (ji). If Di is updated, the corresponding coefficients Zi,i

Classification

After the dictionary D is learned, and the low-rank sparse representations Z of training data X is obtained. We apply a linear classifier for classification. We use the multivariate ridge regression model [30], [42] with quadratic loss and l2-norm regularization to estimate the classifier W^:W^=argminWHWZ22+λW22,which yields the analytic solution W^=HZT(ZZT+λI)1, where H is the class label matrix of X.

For the test images Y, we first compute its lowest-rank representation Ztest of Y with

Experiment

In this section, we evaluate our approach on three face databases: Extended Yale B [43], [44], AR [45], UMIST [46] and ORL [47]. In order to illustrate the advantage of the proposed algorithm, we compare DLR_DL with several other algorithms including SRC [12], K-SVD [2], LC-KSVD [17], DLRD_SR [8], Learning Structured Low-rank Representation (Learning Str LRR) [35]. In this experiments, when compare our approach with SRC and DLRD_SR, we use all the training samples as dictionary and for KSVD,

Conclusions

In this paper, we proposed a Discriminative Low-Rank Dictionary Learning (DLR_DL) approach for face recognition. The DLR_DL aims to learn a discriminative dictionary with low-rank sub dictionaries, which can violate with noise contained in training samples and make the dictionary more pure and compact. The discriminative power of the dictionary comes from discriminating the coding coefficients and optimizing the within-class reconstruction error and minimizing the correlation between-class sub

Acknowledgments

This work was partly supported in part by the National Natural Science Foundation of China under Grant nos. 61375001 and 61473086, partly supported by the open fund of Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education (No. MCCSE2013B01), partly Supported by the Open Project Program of Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University (No. CDLS-2014-04).

HoangVu Nguyen received his B.S. and M.S. degrees at the School of Electrical and Electronics, University of Technical Education HoChiMinh City (HCMUTE), Viet Nam, 1998 and 2009, respectively. Now he is a Ph.D. student in the School of Automation, Southeast University, PR China. His current research interest includes pattern recognition, face recognition and machine learning.

References (47)

  • O. Bryt et al.

    Compression of facial images using the K-SVD algorithm

    J. Vis. Commun. Image Represent.

    (2008)
  • M. Elad et al.

    Image denoising via sparse and redundant representations over learned dictionaries

    Image Process., IEEE Trans.

    (2006)
  • M. Aharon et al.

    K-SVD: an algorithm for designing overcomplete dictionaries for sparse representation

    Signal Process., IEEE Trans.

    (2006)
  • J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, Discriminative learned dictionaries for local image...
  • F. Rodriguez et al.

    Sparse representations for image classification: learning discriminative and reconstructive non-parametric dictionaries

    DTIC Doc.

    (2008)
  • D.-S. Pham and S. Venkatesh, Joint learning and dictionary construction for pattern recognition, in: Proceedings of the...
  • J. Mairal et al.

    Supervised dictionary learning

    Adv. Neural Inform. Process. Syst.

    (2009)
  • Q. Zhang and B. Li, Discriminative K-SVD for dictionary learning in face recognition, in: Proceedings of the IEEE...
  • I. Ramirez, P. Sprechmann, and G. Sapiro, Classification and clustering via dictionary learning with structured...
  • J. Yang, K. Yu, and T. Huang, Supervised translation-invariant sparse coding, in: Proceedings of the IEEE Conference on...
  • M. Yang, L. Zhang, J. Yang, and D. Zhang, Metaface learning for sparse representation based face recognition, in: 17th...
  • J. Wright et al.

    Robust face recognition via sparse representation

    Pattern Anal. Mach. Intell., IEEE Trans.

    (2009)
  • M. Yang, D. Zhang, and J. Yang, Robust sparse coding for face recognition, in: Proceedings of the Computer Vision and...
  • J. Huang, X. Huang, and D. Metaxas, Learning with dynamic group sparsity, in: Proceedings of the IEEE 12th...
  • X.-T. Yuan et al.

    Visual classification with multitask joint sparse representation

    Image Process, IEEE Trans.

    (2012)
  • J. Mairal, F. Bach, J. Ponce, and G. Sapiro, Online dictionary learning for sparse coding, in: Proceedings of the 26th...
  • Z. Jiang, Z. Lin, and L.S. Davis, Learning a discriminative dictionary for sparse coding via label consistent K-SVD,...
  • H. Lee et al.

    Efficient sparse coding algorithms

    Adv. Neural Inform. Process. Syst.

    (2006)
  • J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, Locality-constrained linear coding for image classification, in:...
  • M. Yang, D. Zhang, and X. Feng, Fisher discrimination dictionary learning for sparse representation, in: Proceedings of...
  • C. Studer and R.G. Baraniuk, Dictionary learning from sparsely corrupted or compressed signals, in: Proceedings of the...
  • E.J. Candès, X. Li, Y. Ma, and J. Wright, Robust principal component analysis? ArXiv preprint arXiv:0912.3599,...
  • E.J. Candes and Y. Plan, Matrix completion with noise, ArXiv preprint arXiv:0903.3131,...
  • Cited by (0)

    HoangVu Nguyen received his B.S. and M.S. degrees at the School of Electrical and Electronics, University of Technical Education HoChiMinh City (HCMUTE), Viet Nam, 1998 and 2009, respectively. Now he is a Ph.D. student in the School of Automation, Southeast University, PR China. His current research interest includes pattern recognition, face recognition and machine learning.

    Wankou Yang received his B.S., M.S. and Ph.D. degrees at the School of Computer Science and Technology, Nanjing University of Science and Technology (NUST), PR China, 2002, 2004 and 2009 respectively. Now he is an associate professor in the School of Automation, Southeast University, P.R. China. His research interests include pattern recognition, computer vision.

    Biyun Sheng received the B.S. and M.S. degrees in the School of Electrical Information Engineering, Jiangsu University, China, respectively in 2010 and 2013. Now, she is a Ph.D. student in School of Automation at the Southeast University, Nanjing, China.

    Changyin Sun is a professor in School of Automation at Southeast University, China. He received the M.S. and Ph.D. degrees in Electrical Engineering from Southeast University, Nanjing, China, respectively in 2001 and 2003. His research interests include Intelligent Control, Neural Networks, SVM, Pattern Recognition, Optimal Theory, etc. He has received the First Prize of Nature Science of Ministry of Education, China. He has published more than 40 papers. He is an associate editor of IEEE Transactions on Neural Networks, Neural Processing Letters, International Journal of Swarm Intelligence Research, Recent Patents on Computer Science. He is an IEEE Member.

    View full text