Figures
Abstract
Representation based classification methods, such as Sparse Representation Classification (SRC) and Linear Regression Classification (LRC) have been developed for face recognition problem successfully. However, most of these methods use the original face images without any preprocessing for recognition. Thus, their performances may be affected by some problematic factors (such as illumination and expression variances) in the face images. In order to overcome this limitation, a novel supervised filter learning algorithm is proposed for representation based face recognition in this paper. The underlying idea of our algorithm is to learn a filter so that the within-class representation residuals of the faces' Local Binary Pattern (LBP) features are minimized and the between-class representation residuals of the faces' LBP features are maximized. Therefore, the LBP features of filtered face images are more discriminative for representation based classifiers. Furthermore, we also extend our algorithm for heterogeneous face recognition problem. Extensive experiments are carried out on five databases and the experimental results verify the efficacy of the proposed algorithm.
Citation: Bi C, Zhang L, Qi M, Zheng C, Yi Y, Wang J, et al. (2016) Supervised Filter Learning for Representation Based Face Recognition. PLoS ONE 11(7): e0159084. https://doi.org/10.1371/journal.pone.0159084
Editor: Zhaohong Deng, Jiangnan University, CHINA
Received: January 21, 2016; Accepted: June 27, 2016; Published: July 14, 2016
Copyright: © 2016 Bi et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All databases utilized in this paper are available from figshare.com (https://dx.doi.org/10.6084/m9.figshare.3381562.v1).
Funding: This work is supported by Fund of Jilin Provincial Science & Technology Department (20130206042GX), Young scientific research fund of Jilin province science and technology development project (No. 20130522115JH) and National Natural Science Foundation of China (No. 11271064, 61403078, 61471111). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Automatic face recognition has become a very active topic in computer vision and related research fields [1]. However, face recognition is still a very difficult task in practice due to the following two problematic factors. One is the appearance variations including facial expression, pose, aging, illumination changes, the other is the man-made variations, e.g. the noises from the cameras. The performances of many recognition approaches degrade significantly in these cases.
Recently, the representation based methods have been widely used in face recognition problem. In [2], Wright et al. proposed a sparse representation based classification (SRC) method for face recognition. SRC first sparsely codes a query face image by the original training images, and then the classification is performed by checking which class leads to the minimal representation residual of the query image. Later, Naseem et al. [3] proposed a linear regression based classification (LRC) method based on the assumption that patterns from the same class lie on a linear subspace, so the test image should be well represented as a linear combination of the training images from the same class. The main difference between SRC and LRC is the regularization they employed. That is, SRC utilizes the L1 norm regularization to make the representation coefficients to be sparse, while the L2 norm regularization is adopted in LRC to ensure the learning problem to be well posed. Since the experimental results in [2] and [3] demonstrated that SRC and LRC achieved impressive face recognition performances, the research of representation based face recognition was largely boosted and lots of approaches have been developed [4–8]. However, these representation based methods all utilized the original face image without any preprocessing for classification. Thus, as we have analyzed above, their performances may be affected by the problematic factors in face images.
Nowadays, various feature extraction approaches have been employed for face recognition. Among these approaches, Principal Component Analysis (PCA) [9], Linear Discriminant Analysis (LDA) [10] and their related extensions [11–15]have been well studied and widely utilized to extract low-dimensional features from the high-dimensional face images. However, since some recent studies have shown that high-dimensional face images possibly reside on a nonlinear manifold, many manifold learning methods such as Isometric Feature Mapping (ISOMAP) [16], Local Linear Embedding (LLE) [17], Laplacian Eigenmap (LE) [18] and their extensions have also been proposed for face recognition. Although the aforementioned feature extraction algorithms worked well, they all belong to the subspace based method and can only extract the holistic features of face images, which may lead them to be unstable to local variances such as expression, occlusion, and misalignment [19]. As a result, local descriptors such as Local Binary Pattern (LBP) have attracted more and more attention for their robustness to local distortions [20, 21]. The LBP operator [22] is a texture descriptor which describes the neighboring changes around each pixel. It has been successfully used in face recognition applications due to its invariance to the changes of illumination and expression in face images and computational efficiency. Considering the advantages of LBP in face recognition [23], many LBP variants have been proposed. In LGBP [24], GVLBP [25] and HGPP [26], instead of directly using the pixel intensity to compute the LBP features, multi-scale and multi-orientation Gabor filters were employed for encoding the face images. Then, the LBP histogram was obtained from the encoded images. Zhao et al. first extracted the gradient information from face image using Sobel operator and then applied LBP to the gradient images for feature extraction [27]. The LBP has also been adopted to extract the features for representation based classification techniques. In [28] and [29], some researchers combined LBP with SRC for face recognition. In their methods, the LBP features were first extracted from the face images. Then, the SRC was utilized for classification. Kang et al. employed LBP to extract local features of the face images so that the performance of kernel SRC could be improved [30]. In [31], Lee also used the Gabor-LBP features for face image representation in SRC.
Although the experimental results in [28, 29]indicate that the LBP can improve the performances of representation based face recognition techniques, a main drawback of these methods is that the label information is neglected during the local feature extraction of LBP, which may weaken their discriminative ability. In order to overcome this limitation, Lei et al. proposed an Image Filter Learning (IFL) method for face recognition [19]. In IFL, an image filter which can explore the discriminative information for face representation was first learned. Then, the LBP features were extracted from the filtered face images for recognition. However, IFL learns the discriminative image filter based on Fisher criterion. Thus, it may not be suitable for representation based face recognition methods in which the classification is determined by the representation residuals. Furthermore, the Fisher criterion may also make it not suitable to non-Gaussian distributed face images [32].
In this paper, a new supervised filter learning (SFL) algorithm is proposed to improve the discriminative ability of LBP features for representation based face recognition. Compared with other algorithms, our algorithm possesses two advantages. Firstly, different from LGBP [24], GVLBP [25], HGPP [26] and Sobel-LBP [27] in which the image filters are defined in an ad hoc way, the optimal filter in our algorithm is learned by a supervised data-driven manner. Therefore, the LBP features obtained in our algorithm are more discriminative than them. Secondly, unlike IFL [19] which learns the filter based on Fisher criterion, our proposed SFL is specially designed for representation based face recognition methods. That is, the main difference between IFL and the proposed algorithm is that the filter in IFL is learned by minimizing the within-class scatter and maximizing the between-class scatter of faces’ LBP features, while the filter in our algorithm is learned through reducing the within-class representation residual and enlarging the between-class representation residual of faces' LBP features. As a result, it can be seen from the experimental results on five benchmark face databases (Yale, AR, CMUPIE, LFW and VLNHF) that the performances of our algorithm are better than IFL and some other algorithms for representation based face recognition problem.
The remaining part of the paper is organized as follows: ‘Related Work’ section briefly reviews the LBP and IFL. 'The Proposed Algorithm' section describes the details of our algorithm. Experimental results and analysis are provided in ‘Experiments’ section and ‘Conclusions’ section gives the conclusion of this paper.
Related Work
In this section, two related works including Local Binary Pattern (LBP) [33] and Image Filter Learning (IFL) [19] are briefly reviewed.
Local Binary Pattern
Local Binary Pattern (LBP) [33] was original proposed by Ojala et al. as a powerful technique for texture description. It can efficiently describe the local texture of an image by thresholding each pixel in a 3 × 3 sized neighborhood with the center pixel's value and considering the results as a binary number (see Fig 1 for an illustration). As a result, 256-bin histogram of the LBP labels computed over the image can be used as a texture feature. To describe the image textures at different scales, the LBP was later extended to use different neighborhood sizes [33, 34]. In this way, the values of d points evenly sampled from a circle within an r×r sized neighborhood are compared with the center pixel’s value. Then, the comparison result can also be considered as a binary number (see Fig 2 for an illustration). When the sampled points are not exactly located in the centers of pixels, their values can be estimated by interpolation [33, 34].
(a) d = 8,r = 5, (b) d = 8,r = 7.
Compared with other features, LBP feature has the advantage of invariant to monotone transformation. Thus, it is robust to the illumination and expression changes of face images to some extent and has been widely employed for face recognition. However, one limitation of LBP and its extensions is that the label information of face images is ignored. Therefore, the features extracted by these methods may lack of discrimination.
Discriminant Face Descriptor
For the sake of overcoming the limitation of LBP and improving the discriminative ability of LBP features, Lei et al. proposed a discriminant image filter learning (IFL) for face recognition [19]. The main idea of IFL is to reduce the variances of LBP features of face images from intra person and meanwhile enlarge the margin between LBP features of face images from different persons. To achieve this goal, the label information of face images is utilized to learn a filter in IFL. Then, the LBP operator is applied on the filtered face images for local feature extraction. Specifically, let I denote an input face image and its filtered image is denoted as f(I). Considering the sampling strategy of LBP, IFL first defines pixel difference vector (PDV) as:
(1)
where f(I)p is the pixel value of filtered face image at position p, {p1, p2, ⋯, pd}∈Neighbor(p) and d is the number of sampled points. Then, in order to make sure that the PDVs of filtered face images from the same person are similar and the PDVs of filtered face images from different persons are distant. The Fisher criterion is adopted as
(2)
where Sb and Sw are the between-class and within-class class scatters, which can be computed as
(3)
(4)
where L is the number of classes and Ci is the number of samples in the i-th class.
is the PDV set from the j-th image of class i, and N is the number of PDV for each filtered face image.
and df(m) = [df(m)1, df(m)2, ⋯, df(m)N] are augmented vectors by concatenating mean vectors over different positions.
is the mean vector of PDVs at position p of the filtered face images from the i-th class and df(m)P is the total mean vector of PDVs at position p over the sample set. In IFL, the image filter vector is set to be w, and the value of filtered image at position p can be represented as f(I)P = wTIp, where Ip denotes the patch vector centered at position p of the original face image. Therefore, the filter w can be learned by
(5)
where
and
are the between-class and within-class scatters of PDVs from the original input face images. For more details about the IFL, the readers can refer to [19].
The Proposed Algorithm
Supervised Filter Learning
As shown in the previous section, IFL utilized the Fisher criterion to learn an optimal filter which improved the discriminative ability of LBP features extracted from filtered images. Therefore, like other methods based on Fisher criterion (such as LDA), it may only be suitable for the case in which the samples of each class are approximately Gaussian distributed [32]. However, this property is not always satisfied in face recognition problem [35]. Furthermore, the Fisher criterion is also not suitable for the representation based classification methods which have been proved to be effective for face recognition tasks. In order to overcome these limitations, we propose a new supervised filter learning (SFL) algorithm to improve the discriminative ability of LBP features for representation based face recognition.
Formally, let denote a set of training face images from L classes (each class possesses Ci samples, i = 1, …, L). Similar to IFL, we suppose that the filtered images are f(T) = [f(T1), f(T2), ⋯, f(TN)]. Since the proposed algorithm also applies LBP operator on the filtered image, we define the pixel difference vector (PDV) as:
(6)
where f(Ti)p is the pixel value of filtered face image f(Ti) at position p, {p1, p2, ⋯, pd} ∈ Neighbor(p) and d is the number of sampled points.
Different from IFL which maximize the ratio of between-class scatter to the within-class scatter of LBP features extracted from the filtered face images, the aim of the proposed algorithm is to benefit the representation based face recognition methods. That is, our algorithm want to learn a filter so that after the image filtering, the LBP feature of a face image can be accurately represented by those from the same person and cannot be represented by those of different persons. To achieve this goal, we need to reduce the within-class representation residual and enlarge the between-class representation residual of the filtered images' PDVs. Suppose that df(Tij)p is the p-th PDV of the j-th face image from the i-th class, its within-class representation residual can be obtained as:
(7)
where
is a matrix formed by the p-th PDVs of the other filtered face images from the i-th class and
is the vector of within-class representation coefficients for df(Tij)p, which can be estimated using the least-squares algorithm [28] as:
(8)
where
is the transpose of
.
Considering all the PDVs, we can get the total within-class representation residual as
(9)
Similarly, the between-class representation residual of df(Tij)p can be formulated as
(10)
where
is a matrix formed by the p-th PDVs of the filtered face images do not belong to the i-th class and
is the vector of between-class representation coefficients for df(Tij)p, which can also be estimated by least-squares algorithm as:
(11)
where
is the transpose of
.
Then, the total between-class representation residual can be obtained as
(12)
Now, through combining Eqs (10) and (12), the objective function of our proposed filter learning algorithm is
(13)
From the definitions of Rw and Rb, it can be found that Eq (13) will incur heavy penalties if the within-class residual of the filtered images' PDVs is large and the between-class residual of the filtered images' PDVs is small. Thus, minimizing Eq (13) could ensure that the LBP features extracted from a filtered face image can only be well represented by those from the same class but cannot be represented by those from different classes. In this study, we suppose that the image filter with the size of S × S can be concatenated into a vector (F = S × S). Then, the value of a filtered image f(Tij) at position p can be denoted as
, where
is a vector concatenated by the patch centered at position p of image Tij. Analogously, the PDV at position p of a filtered image can also be denoted as
, where
is the PDV at position p of the unfiltered image Tij. Through substituting df(Tij)p into Eqs (10) and (12), Eq (13) can be converted to
(14)
Let and
, Eq (14) is reduced to
(15)
From Eq (15), it is clear that both the matrix and
are symmetric and positive semi-definite. As a result, the optimal filter (i.e. ω) that minimizing the objective function of our algorithm can be obtained by solving the generalized eigenvalue problem
with its smallest eigenvalue.
After the filter ω has been learned, we can convert it into the matrix form with the size of S × S and employ it to filter the training face images. Then, the LBP features are extracted from the filtered images and the representation based classification methods (such as SRC and LSR) can be utilized for recognition.
Extended SFL for Heterogeneous Face Recognition
Nowadays, heterogeneous face image recognition has attracted more and more attentions due to its widely applications in video surveillance and law enforcement. According to some studies [19, 36], the heterogeneous faces can be defined as faces which are captured in different environments or different devices. For instance, the face images captured by visible light and near-infrared imaging devices can be regarded as heterogeneous faces.
In this section, we extended the proposed SFL for heterogeneous face recognition problem. Similar to the SFL for homogeneous face images, the aim of extended SFL is to learn a filter to reduce the within-class representation residual of faces’ LBP features for heterogeneous images from the same person and enlarge the between-class representation residual of faces' LBP features for heterogeneous images from the different persons. Suppose and
are two heterogeneous image sets (e.g. images captured by visible light and near-infrared imaging devices), and the filtered images of them are
and
, respectively. Let
and
be the p-th PDVs of the j-th faces from the i-th class in two heterogeneous image sets. In order to make sure that the LBP features of face images can be well represented by those from the same person, we need to minimize the following within-class representation residual:
(16)
where
and
are the homogeneous within-class representation residuals which can be obtained by Eq (9).
and
are the within-class representation residuals between heterogeneous images, which can be defined as:
(17)
and
(18)
where
and
are the matrices formed by the p-th PDVs of face images from the i-th class in image sets TM and TV.
and
are the heterogeneous representation coefficients of
and
, which can be obtained by least-squares algorithm similar to Eq (8).
Analogically, to ensure that the LBP features of face images cannot be represented by those from different persons, the following between-class representation residual should be maximized:
(19)
where
and
are the homogeneous between-class representation residuals obtained by Eq (12).
and
are the heterogeneous between-class representation residuals defined as:
(20)
and
(21)
where
and
are the matrices formed by the p-th PDVs of face images do not belong to the i-th class in image sets TM and TV.
and
are the heterogeneous representation coefficients of
and
.
Through combining Eqs (16) and (19) together, we can obtain the objective function of extended SFL for heterogeneous face recognition as
(22)
Similar to ‘Supervised Filter Learning’ section, we also suppose that that the image filter with the size of S × S can be concatenated into a vector (F = S × S). Then, we have
and
, where
and
are the PDVs at position p of the unfiltered images
and
. Now, by substituting
and
into Eqs (16) and (19), these two equations can be converted to
(23)
(24)
After a series of deductions, Eq (22) can be reduced to
(25)
where
(26)
(27)
Therefore, the optimal filter ω that minimizing the objective function of extended SFL in Eq (25) can be obtained by solving the generalized eigenvalue problem Kwω = λKbω with its smallest eigenvalue. After the filter learning, ω can be converted into its matrix form to filter the heterogeneous face images in TM and TV. Then, the SRC or LRC can be adopted for recognition.
Experiments
In this section, the performance of the proposed algorithm is tested and compared with other related algorithms such as LBP [33], LGBP [24], GVLBP [25], IFL-LBP [19], DSNPE [37], MNSMC [38]and UDSPP [39]. Among these algorithms, LBP, LGBP, GVLBP and IFL-LBP are LBP based methods, while DSNPE, MNSMC and UDSPP are recently proposed subspace based methods for representation based face recognition. Here, five benchmark face databases including Yale [40], AR [41], CMU PIE [42], LFW [43] and VLNHF [44] are employed. The proposed algorithm and other approaches used for comparison are all implemented in Matlab and executed on a computer with Intel Core i3-2100 CPU at 3.2 GHz and 8 GB physical memory.
Data Description
The Yale face database [40] contains 165 grayscale images of 15 individuals. Thereare11 images per subject, and the images demonstrate variations in facial expression (normal, sad, happy, sleepy, surprised, and wink), lighting condition (left-light, center-light, right-light), and with/without glasses. In our experiment, 6 images of each person are randomly selected for training and the rest images are used for testing.
The AR face database [41] consists of more than 4000 frontal images from 126 subjects including 70 males and 56 females. The images were taken in two sessions separated by two weeks with expression (neutral, smile, anger and scream) and occlusion (sunglass and scarf) variations. In this experiment, we choose a subset which contains 50 males and 50 females. For each subject, 14 images with only illumination and expression changes are selected. We randomly select 7 images from each person for training, and remaining images are used for testing.
The CMU PIE face database [42] includes 68 subjects with 41368 face images as a whole, each subject contains 13 different poses, 43 different illumination conditions, and 4 different expressions. In our experiment, 24 face images of each individual are used. For this database, we randomly select 12 images of each person to form the training set and the rest images are utilized for testing.
The LFW database [43] is a large scale database which contains 13,233 face images of 5,749 different individuals. Since all the samples were taken from the real world in an unconstrained environment, the expression, pose, illumination, occlusions and alignment of face images are very variable in this database. In our study, a subset which contains 1580 face images of 158 individuals from the LFW database is employed. We randomly select 7 images from each person for training, and remaining images are used for testing.
The Visible Light and Near-infrared Human Face (VLNHF) database [44] is a heterogeneous face image database which consists of two datasets (Lab1 and Lab2). The Lab1 dataset simultaneously contains visible light images and near-infrared images of 50 persons. Each person has 10 visible light images and 10 near-infrared images. The Lab2 dataset also contains visible light images and near-infrared images of 50 subjects. Each subject provides twenty visible light face images and the same number of near-infrared face images. These images were acquired under four different illumination conditions, and also have variation in facial expression and pose. In the experiment, 7 visible light images and 7 near-infrared images of each person are randomly selected for training in Lab1 dataset, and 12 visible light images and 12 near-infrared images of each person are randomly selected for training in Lab2 dataset. The rest images are used for testing.
In our recognition experiment, all images are manually aligned, cropped, and then resized to the resolution of 66×66, the random training sample selection are repeated 10 times for all databases and the averaged recognition accuracies are reported in the next subsection.
Results and Discussions
In the proposed algorithm and IFL-LBP, the image filter size S and neighborhood size r of LBP will affect their performances. According to [36], we empirically set S and r to be the same value and tune the value from{3, 5, 7}. The number of sampled points is set as d = 8 for all LBP based algorithms so that 256 dimension LBP features are extracted. For DSNPE, MNSMC and UDSPP, in order to fairly compare them with the LBP based algorithms, the dimension of subspace in these three algorithms are also set as 256. Two well known representation based classifiers, i.e., SRC [2] and LRC [3] are adopted for recognition in our study.
Homogeneous face recognition.
The recognition performances of various approaches on different homogeneous face databases can be seen in Tables 1–4. From these tables, the following points can be observed. Firstly, it can be found that LBP extracts the local texture features directly from the original face images, so its performances are inferior to other algorithms in most cases. Secondly, we can see that the performances of LGBP and GVLBP are better than LBP on AR, CMU PIE and LFW databases. This is because that LGBP and GVLBP extract the LBP features from the images after multi-scale and multi-orientation Gabor filtering, which could eliminate the influences of illumination and expression changes in the face images to some extent. However, we can also observe that LBP outperforms the LGBP and GVLBP on Yale database. The reason to this phenomenon may be that the number of individuals in Yale is much less than other three databases. Thus, the dimension of LBP features obtained from multi-scale and multi-orientation Gabor filtered face images is much higher than the number of training instances. This “small sample size” problem will weaken the performances of classifiers [45]. Thirdly, since IFL-LBP learns the filter in a supervised manner, its recognition results are better than other LBP based algorithms. Fourthly, we can find that the performances of subspace based algorithms (i.e. DSNPE, MNSMC and UDSPP) are better than LBP, LGBP, GVLBP and IFL-LBP in some cases. This is because these three algorithms are all designed for representation based face recognition. Nevertheless, since the subspace based algorithms only extract holistic features from the face images, their recognition results are still worse than our algorithm. At last, it can be seen that the proposed algorithm outperforms IFL-LBP and other algorithms on all databases. This is due to that the filter in our algorithm is learned based on the representation residual rather than Fisher criterion, which makes the LBP features extracted from the filtered images more suitable for the representation based classifiers. Besides the representation based classifiers, we also compare the performances of our SFL-LBP with IFL-LBP using Nearest Neighbor classifier. From the experimental results in Table 5, it can be found that the proposed algorithm outperforms IFL-LBP in most cases. This is because the Fisher criterion utilized in IFL cannot work well when the input training samples are not Gaussian distributed.
Then, the performances of our algorithm under different filter and neighborhood sizes are compared. From the experimental results in Tables 1–4, it can be found that the proposed SFL-LBP achieves better performances than IFL-LBP when their parameters are set as the same value. Moreover, we can also see that the values of parameters S and r have important effect on the performances of both IFL-LBP and SFL-LBP. However, given the standard deviation, the differences among the recognition results of our algorithm under various parameter values are less than IFL-LBP (especially on AR and CMU PIE databases). This indicates the proposed algorithm is less sensitive to the parameters when they are set as appropriate values.
Next, the Cumulative Match Characteristic (CMC) curve is used in our experiment to further compare the performances of IFL-LBP and our algorithm. From the CMC curves in Figs 3 and 4, it can be observed that our algorithm outperforms IFL-LBP nearly at all ranks, which demonstrates the advantage and robust of our algorithm for representation based face recognition tasks.
Heterogeneous face recognition.
In this subsection, the performance of the proposed algorithm for heterogeneous face recognition are validated and compared with IFL. The average recognition results obtained by IFL and SFL are tabulated in Tables 6 and 7. From these tables, we can find that the proposed SFL outperforms IFL, which is consistent with the experimental results in ‘Homogeneous Face Recognition’ section. Furthermore, from CMC curves in Figs 5 and 6, the superior of our SFL for heterogeneous face recognition task is also verified.
Statistical test.
In this subsection, the one-tailed Wilcoxon rank sum test is utilized to verify whether the performance of our algorithm is significantly better than the other algorithms. In this test, the null hypothesis is that the proposed SFL-LBP makes no difference when compared to other algorithms, and the alternative hypothesis is that SFL-LBP makes an improvement when compared to other algorithms. For instance, if we want to compare the performance of our algorithm with that of LBP (SFL-LBP vs. LBP), the null and alternative hypotheses can be defined as H0: MSFL-LBP = MLBP and H1: MSFL-LBP> MLBP, where MSFL-LBP and MLBP are the medians of the recognition rates obtained by SFL-LBP and LBP on all face databases. In our experiments, the significance level is set to 0.05. From the test results in Table 8, it can be found that the p-values obtained by all pairwise Wilcoxon rank sum tests are less than the significance level, which indicates that the null hypotheses are rejected in all pairwise tests and the proposed algorithm significantly outperforms the other algorithms.
Conclusions
This paper presents a filter learning algorithm for representation based face recognition. Due to the objective function of our proposed algorithm is specially designed to reduce the within-class representation residual and enlarge the between-class representation residual of faces' local descriptors, it is more suitable for the representation based classifiers than other algorithms. In the experiments, five public face databases are utilized to evaluate our algorithm. Through comparing our algorithm with other state-of-the-art algorithms using two well-known representation based classifiers, the effectiveness and advantage of our algorithm are demonstrated.
Acknowledgments
This work is supported by Fund of Jilin Provincial Science & Technology Department (20130206042GX), Young scientific research fund of Jilin province science and technology development project (No. 20130522115JH) and National Natural Science Foundation of China (No. 11271064, 61403078, 61471111). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Author Contributions
Conceived and designed the experiments: CB JW BZ. Performed the experiments: CB MQ CZ. Analyzed the data: JW BZ. Contributed reagents/materials/analysis tools: LZ YY. Wrote the paper: CB JW.
References
- 1.
Jain AK, Li SZ. Handbook of face recognition: Springer; 2005.
- 2. Wright J, Yang AY, Ganesh A, Sastry SS, Ma Y. Robust face recognition via sparse representation. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2009;31(2):210–27.
- 3. Naseem I, Togneri R, Bennamoun M. Linear regression for face recognition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2010;32(11):2106–12.
- 4.
Gao S, Tsang IW-H, Chia L-T. Kernel sparse representation for image classification and face recognition. Computer Vision–ECCV 2010: Springer; 2010. p. 1–14.
- 5.
Wang J, Yang J, Yu K, Lv F, Huang T, Gong Y. Locality-constrained linear coding for image classification. Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on; 2010: IEEE.
- 6.
Yang M, Zhang L, Yang J, Zhang D. Robust sparse coding for face recognition. Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on; 2011: IEEE.
- 7. Deng W, Hu J, Guo J. Extended SRC: Undersampled face recognition via intraclass variant dictionary. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2012;34(9):1864–70.
- 8. Mi J-X, Liu J-X. Face recognition using sparse representation-based classification on k-nearest subspace. PloS one. 2013;8(3):e59430. pmid:23555671
- 9. Turk M, Pentland A. Eigenfaces for recognition. Journal of cognitive neuroscience. 1991;3(1):71–86. pmid:23964806
- 10. Scholkopft B, Mullert K-R. Fisher discriminant analysis with kernels. Neural networks for signal processing IX. 1999;1:1.
- 11. Yang J, Zhang D, Frangi AF, Yang J-y. Two-dimensional PCA: a new approach to appearance-based face representation and recognition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2004;26(1):131–7.
- 12.
Pentland A, Moghaddam B, Starner T. View-based and modular eigenspaces for face recognition. Computer Vision and Pattern Recognition, 1994 Proceedings CVPR'94, 1994 IEEE Computer Society Conference on; 1994: IEEE.
- 13. Howland P, Park H. Generalizing discriminant analysis using the generalized singular value decomposition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2004;26(8):995–1006.
- 14. Ye J, Li Q. A two-stage linear discriminant analysis via QR-decomposition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2005;27(6):929–41.
- 15. Belhumeur PN, Hespanha JP, Kriegman DJ. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 1997;19(7):711–20.
- 16. Roweis ST, Saul LK. Nonlinear dimensionality reduction by locally linear embedding. Science. 2000;290(5500):2323–6. pmid:11125150
- 17. Tenenbaum JB, De Silva V, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science. 2000;290(5500):2319–23. pmid:11125149
- 18. Belkin M, Niyogi P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural computation. 2003;15(6):1373–96.
- 19.
Lei Z, Yi D, Li SZ, Discriminant image filter learning for face recognition with local binary pattern like representation. Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on; 2012: IEEE.
- 20. Mikolajczyk K, Schmid C. A performance evaluation of local descriptors. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2005;27(10):1615–30.
- 21. Penev PS, Atick JJ. Local feature analysis: A general statistical theory for object representation. Network: computation in neural systems. 1996;7(3):477–500.
- 22. Ojala T, Pietikäinen M, Harwood D. A comparative study of texture measures with classification based on featured distributions. Pattern recognition. 1996;29(1):51–9.
- 23.
Shan S, Gao W, Chang Y, Cao B, Yang P. Review the strength of gabor features for face recognition from the angle of its robustness to mis-alignment. Pattern Recognition, 2004 ICPR 2004 Proceedings of the 17th International Conference on; 2004: IEEE.
- 24.
Zhang W, Shan S, Gao W, Chen X, Zhang H, Local gabor binary pattern histogram sequence (lgbphs): A novel non-statistical model for face representation and recognition. Computer Vision, 2005 ICCV 2005 Tenth IEEE International Conference on; 2005: IEEE.
- 25. Lei Z, Liao S, Pietikäinen M, Li SZ. Face recognition by exploring information jointly in space, scale and orientation. Image Processing, IEEE Transactions on. 2011;20(1):247–56.
- 26. Zhang B, Shan S, Chen X, Gao W. Histogram of gabor phase patterns (hgpp): A novel object representation approach for face recognition. Image Processing, IEEE Transactions on. 2007;16(1):57–68.
- 27.
Zhao S, Gao Y, Zhang B, editors. Sobel-lbp. Image Processing, 2008 ICIP 2008 15th IEEE International Conference on; 2008: IEEE.
- 28.
Min R, Dugelay J-L. Improved combination of LBP and sparse representation based classification (SRC) for face recognition. Multimedia and Expo (ICME), 2011 IEEE International Conference on; 2011: IEEE.
- 29.
Huang M-W, Wang Z-w, Ying Z-L. A new method for facial expression recognition based on sparse representation plus LBP. Image and Signal Processing (CISP), 2010 3rd International Congress on; 2010: IEEE.
- 30.
Kang C, Liao S, Xiang S, Pan C. Kernel sparse representation with local patterns for face recognition. Image Processing (ICIP), 2011 18th IEEE International Conference on; 2011: IEEE.
- 31.
Lee H, Chung Y, Kim J, Park D. Face image retrieval using sparse representation classifier with gabor-lbp histogram. Information Security Applications: Springer; 2011. p. 273–80.
- 32. Yan S, Xu D, Zhang B, Zhang H-J, Yang Q, Lin S. Graph embedding and extensions: a general framework for dimensionality reduction. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2007;29(1):40–51.
- 33. Ojala T, Pietikäinen M, Mäenpää T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2002;24(7):971–87.
- 34. Ahonen T, Hadid A, Pietikainen M. Face description with local binary patterns: Application to face recognition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2006;28(12):2037–41.
- 35. Zhu M, Martinez AM. Subclass discriminant analysis. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2006;28(8):1274–86.
- 36. Lei Z, Pietikainen M, Li SZ. Learning discriminant face descriptor. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2014;36(2):289–302.
- 37. Lu G-F, Jin Z, Zou J. Face recognition using discriminant sparsity neighborhood preserving embedding. Knowledge-Based Systems. 2012;31:119–27.
- 38. Chen Y, Li Z, Jin Z. Feature extraction based on maximum nearest subspace margin criterion. Neural processing letters. 2013;37(3):355–75.
- 39.
Chen Z, Huang W, Lv Z. Towards a face recognition method based on uncorrelated discriminant sparse preserving projection. Multimedia Tools and Applications. 2015:1–15.
- 40.
Georghiades A. Yale face database. Center for computational Vision and Control at Yale University. Available: http://cvc.yale.edu/projects/yalefaces/yalefa. 1997.
- 41.
Martinez AM. The AR face database. CVC Technical Report. 1998;24.
- 42.
Sim T, Baker S, Bsat M. The CMU pose, illumination, and expression (PIE) database. Automatic Face and Gesture Recognition, 2002 Proceedings Fifth IEEE International Conference on; 2002: IEEE.
- 43.
Huang GB, Ramesh M, Berg T. Learned-Miller E. Labeled faces in the wild: A database for studying face recognition in unconstrained environments. Technical Report 07–49, University of Massachusetts, Amherst, 2007.
- 44. Xu Y, Zhong A, Yang J, Zhang D. Bimodal biometrics based on a representation and recognition approach. Optical Engineering. 2011;50(3):037202–7.
- 45. Raudys SJ, Jain AK. Small sample size effects in statistical pattern recognition: Recommendations for practitioners. IEEE Transactions on Pattern Analysis & Machine Intelligence. 1991;(3):252–64.