Open Access
20 March 2019 L1-2D2PCANet: a deep learning network for face recognition
Yun-Kun Li, Xiao-Jun Wu, Josef Kittler
Author Affiliations +
Abstract
We propose a deep learning network L1-2D2PCANet for face recognition, which is based on L1-norm-based two-dimensional principal component analysis (L1-2DPCA). In our network, the role of L1-2DPCA is to learn the filters of multiple convolution layers. After the convolution layers, we deploy binary hashing and blockwise histogram for pooling. We test our network on some benchmark facial datasets, including Yale, AR face database, extended Yale B, labeled faces in the wild-aligned, and Face Recognition Technology database with the convolution neural network, PCANet, 2DPCANet, and L1-PCANet as comparison. The results show that the recognition performance of L1-2D2PCANet in all tests is better than baseline networks, especially when there are outliers in the test data. Owing to the L1-norm, L1-2D2PCANet is robust to outliers and changes of the training images.

1.

Introduction

In pattern recognition and computer vision, face recognition is a very important research field.16 Due to the complexity of facial features and the difficulty of manual feature selection,1,5,6 it is commonly agreed that the best features can be obtained by using unsupervised feature extraction methods.35

Recently, with Google Alpha Go Zero defeating many Go masters, deep learning has received intensive attentions.7,8 As a classical deep learning model, convolution neural networks (CNNs) with convolution and pooling layers have achieved astonishing results in many image recognition tasks, reaching an unprecedented accuracy.9,10 However, CNN still has many shortcomings. During the process of training a CNN model, researchers need to obtain a huge number of parameters, which leads to high computational cost.11

To solve this problem, researchers are committed to finding a simple CNN model that requires a small number of parameters. Chan et al.12 proposed PCANet, which is a simple deep learning network based on unsupervised learning. PCANet uses PCA to learn the filters and deploys simple binary hashing and block histogram for indexing and pooling. Unlike other CNNs that learn filters by backpropagation, PCANet learns filters using the PCA method. Thus, PCANet requires less computational cost, less time, and storage space. The experimental results show the astonishing performance of PCANet.

The PCA method used by PCANet is based on one-dimensional (1-D) vectors. Before deploying PCA, we need to convert two-dimensional (2-D) image matrices into 1-D vectors, which will cause two major problems: (1) Some spatial information of image is implied in the 2-D structure of the image.13,14 Obviously, the intrinsic information is discarded when the image matrix is converted into 1-D vector.13,15 (2) The long 1-D vector leads to the requirement of large computational time and storage space in computing the eigenvectors. To solve these problems, Yu et al.16 proposed 2-D principal component analysis network (2DPCANet), which replaces PCA with 2DPCA.15,1719 And Tian et al.20 proposed multiple scales principal component analysis network (MS-PCANet).

However, both PCA and 2DPCA are based on L2-norm method. It is well known that the methods based on L2-norm are sensitive to outliers so that data with outliers can totally ruin the results from the desired methods.5,21,22 To solve this problem, Kwak23 proposed a PCA method based on L1-norm. L1-norm is widely considered to be more robust to outliers.21,24 L1-PCA adopts the L1-norm for measuring the reconstruction error. On this basis, Xuelong et al.14 proposed L1-norm-based 2DPCA.

In this paper, L1-norm was introduced into PCANet to get L1-PCANet. Then, we generalize L1-PCANet to L1-2D2PCANet, which shares the same structure with 2DPCANet to generate the feature of input data but L1-2D2PCANet learns filters by L1-2DPCA. In addition, we use support vector machine (SVM) as classifiers for the features generated by the networks. To test the performance of L1-2D2PCANet, we compare it with other three networks (PCANet, 2DPCANet, and L1-PCANet) on Yale, AR,25 extended Yale B,26 labeled faces in the wild-aligned (LFW-a),27 and Face Recognition Technology database (FERET)28 face databases.

The rest of paper is organized as follows. Sections 2.1 and 2.2 review related work on L1-PCA and L1-2DPCA. L1-PCANet and L1-2D2PCANet are given in Sec. 2.3. Section 3 reports the detail of experiments. Section 4 reports the results and the analysis of the experiments and Sec. 5 concludes this paper.

2.

Materials and Methods

2.1.

L1-Norm-Based PCA

The proposed L1-PCANet is based on L1-PCA.21,23 L1-PCA is considered as the simplest and most efficient among many models of L1-norm-based PCA. Let X=[x1,x2,,xN]RD×N, with xi=matD(Ii)RD×1(i=1,2,,N). The matD(I) is a function that maps a matrix IRm×n to a vector vRD×1 and D=m×n. Suppose wRD×1 be the principal vector to be obtained. Here, we set the number of principal vectors to one to simplify the procedure. The objective of L1-PCA is to maximize the L1-norm variance in the feature space and the successive greedy solutions are expected to provide a good approximation as the following:

Eq. (1)

f(w)=wTX1=i=1N|wTxi|,subject to  w2=1,
where · denotes L2-norm and |·| denotes L1-norm.

To solve the computational problems posed by the symbol of absolute value, we introduce a polarity parameter pi in Eq. (1):

Eq. (2)

pi={1,whenwTxi01,when  wTxi<0.

By introducing pi, Eq. (1) can be rewritten as follows:

Eq. (3)

f(w)=i=1NpiwTxi.

The process of maximization is achieved by Algorithm 1. Here, t denotes the number of iterations and w(t) and pi(t) denote w and pi during iteration t.

Algorithm 1

L1-PCA method.

Input:
 • training set: X=[x1,x2,,xN]RD×N
Output:
 • filters w*
1: set w(0)=0 and t=0
2: For all i {1,2,,N}, calculate pi(t) by using Eq. (2)
3: Let t=t+1 and w(t)=i=1Npi(t1)xi. Then let w(t)=w(t)/w(t)2
4: If w(t)w(t1), go back to Step 2. Otherwise, set w*=w(t) and stop.

By the above algorithm, we can obtain the first principal vector w1*. To compute wk*(k>1), we have to update the training data as follows:

Eq. (4)

xik=xik1xik1(wk1*wk1*T).

2.2.

L1-Norm-Based 2DPCA

In this section, we extend L1-PCA to L1-2DPCA.14 As mentioned above, 2DPCA computes eigenvectors with 2-D input. Suppose Ii(i=1,2,,N) denote N input training images and D=m×n being the image size. Let wRw×1 be the first principal component to be learned. Let X=[x1,x2,,xN]RD×N, with xi=[xi1,xi2,,xih]TRh×w(i=1,2,,N). Note, xijR1×w The objective of L1-PCA is to maximize the L1-norm variance in feature space as follows:

Eq. (5)

f(w)=Xw1=i=1Nj=1h|xijw|,subject to  w2=1.

The polarity parameter pij can be computed as follows:

Eq. (6)

pij={1,when  xijw01,when  xijw<0.

The process of maximization is achieved by Algorithm 2. To compute wk*(k>1), we have to update the training data as follows:

Eq. (7)

xijk=xijk1xijk1(wk1*wk1*T).

Algorithm 2

L1-2DPCA method.

Input:
 • training set: X=[x1,x2,,xN]RD×N
Output:
 •filters w*
1: Set w(0)=0 and t=0
2: For all i{1,2,,N} and j{1,2,,h}, calculate pij(t) by using Eq. (6).
3: Let t=t+1 and w(t)=i=1Nj=1hpij(t1)xij. Then we initialize w(t)=w(t)/w(t)2
4: If w(t)w(t1), go back to Step 2. Otherwise, set w*=w(t) and stop.

At this point, we can find that the difference between L1-PCA and L1-2DPCA is that L1-PCA converts an image matrix into a vector, however, L1-2DPCA directly uses each row in the original image matrix as a vector.

2.3.

Proposed Method

2.3.1.

L1-PCANet

In this section, we propose a PCA-based deep learning network, L1-PCANet. To overcome the sensitivity to outliers in PCANet due to the use of L2-norm, we use the L1-PCA rather than the PCA to learn the filters. L1-PCANet and PCANet12 share the same network architecture, which is shown in Fig. 1.

Fig. 1

The illustration of two-layer L1-PCANet.

JEI_28_2_023016_f001.png

Suppose there are N training images Ii(i=1,2,,N) of size m×n, and we get D=m×n patches of size k×k around each pixel in Ii. Then, we take all overlapping patches and map them into vectors:

Eq. (8)

[xi,1,xi,2,,xi,mn]Rk2×mn.

And we remove the patch mean from each patch and obtain as follows:

Eq. (9)

X¯=[x¯i,1,x¯i,2,,x¯i,mn]Rk2×mn.

For all input images, we construct the same matrix and combine them into one matrix to obtain as follows:

Eq. (10)

X=[X¯1,X¯2,,X¯N]Rk2×Nmn.

Then, we use L1-PCA mentioned above to learn the filters in stage 1. The filter we want to find is wRk2×1. We take X as the input data of L1-PCA. Assuming that the number of filters in stage 1 is L1, we can obtain the first stage filters {w1*,,wL1*} by repeatedly calling Algorithm 1. The L1-PCA filters of stage 1 are expressed as follows:

Eq. (11)

Wp1=matk,k(wp*)Rk×k,
where p=1,2,,L1.

The output of stage 1 can expressed as follows:

Eq. (12)

Oip=Ii*Wp1,i=1,2,,N,
where * denotes 2-D convolution. We set the boundary of the input image to zero-padding to make sure that Oip is of the same size as Ii. We can get the filters of the second and subsequent layers by simply repeating the process of the first layer design. The pooling layer of L1-PCANet is almost the same as the pooling layer of L1-2D2PCANet.

2.3.2.

L1-2D2PCANet

In this section, we generalize L1-PCANet to L1-2D2PCANet, which shares the same network with 2DPCANet,16 as shown in Fig. 2.

Fig. 2

The illustration of two-layer L1-2D2PCANet.

JEI_28_2_023016_f002.png

First stage of L1-2D2PCANet

Let all the assumptions be the same as in Section III. We get all the overlapping patches:

Eq. (13)

xi,jRk×k,j=1,2,,mn,
and subtract the patch mean from each of them and we form a matrix:

Eq. (14)

X¯x,i=[x¯i,1,x¯i,2,,x¯i,mn]Rk×kmn.

And we use the transpose of xi,j to form matrix:

Eq. (15)

X¯y,i=[x¯i,1T,x¯i,2T,,x¯i,mnT]Rk×kmn.

For all input images, we construct the matrix by the same way and put them into one matrix, we can obtain as follows:

Eq. (16)

Xx=[X¯x,1,X¯x,2,,X¯x,N]Rk×Nkmn,

Eq. (17)

Xy=[X¯y,1,X¯y,2,,X¯y,N]Rk×Nkmn.

Then, we use L1-2DPCA mentioned above to learn the filters in stage 1. We want to obtain filters wx,p*Rk×1 and wy,p*Rk×1, where p=1,2,,L1. Xx and Xy are the input data for L1-2DPCA. Assuming that the number of filters in stage 1 is L1, the first stage filters {wx,1*,,wx,L1*} and {wy,1*,,wy,L1*} are obtained by repeatedly calling Algorithm 2.

The filters we need in stage 1 can finally be expressed as follows:

Eq. (18)

Wp1=wx,p*×wy,p*TRk×k.

The output of stage 1 will be

Eq. (19)

Oip=Ii*Wp1,i=1,2,,N.

Second stage of L1-2D2PCANet

Like in the first stage, we can start with the overlapping patches of Oip and remove the patch mean from each patch. Then, we have

Eq. (20)

Yx,ip=[y¯i,p,1,,y¯i,p,mn]Rk×kmn,

Eq. (21)

Yy,ip=[y¯i,p,1T,,y¯i,p,mnT]Rk×kmn.

Further, we define the matrix that collects all the patches without the patch mean of the k’th output Oik being removed as

Eq. (22)

Yxp=[Yx,1m,Yx,2m,,Yx,Nm]Rk×Nkmn,

Eq. (23)

Yyp=[Yy,1p,Yy,2p,,Yy,Np]Rk×Nkmn.

Finally, the input of the second stage is obtained by concatenating Yxp and Yyp for all L1 filters:

Eq. (24)

Yx=[Yx1,Yx2,,YxL1]Rk×L1Nkmn,

Eq. (25)

Yy=[Yy1,Yy2,,YyL1]Rk×L1Nkmn.

We take Yx and Yy as the input data of L1-2DPCA. Assuming that the number of filters in stage 2 is L2, we design the second stage filters {wx,1*,,wx,L2*} and {wy,1*,,wy,L2*} by repeatedly calling Algorithm 2. The L1-2DPCA filters of stage 2 are expressed as follows:

Eq. (26)

Wq2=wx,q*×wy,q*TRk×k,
where q=1,2,,  L2.

Therefore, we have L2 outputs for each output Oip of stage 1:

Eq. (27)

Biq={Oip*Wq2},l=1,2,,L2.

Note that the number of outputs of stage 2 is L1L2.

Pooling stage

First, we use a Heaviside-like step function to binarize the output of stage 2. The function H(·) can be expressed as follows:

Eq. (28)

H(x)={0,x<01,x0.

Each pixel is encoded by the following function:

Eq. (29)

Tim=lL22l1H(Biq),
where Tim is an integer of range [0,2L21].

Second, we divide Tim into B blocks. Then, we make a histogram of all blocks of Tim with 2L2 values and concatenate all the histogram of B blocks into one vector hist(Tim). In this way, we obtain L1 histograms and we put them into a vector:

Eq. (30)

fi=[hist(Ti1),,hist(TiL2)]R2L2L1B×1.

Using the L1-2DPCA model described above, we can transform an input image into a feature vector as the output of L1-2D2PCANet.

3.

Experiments

In this section, we evaluate the performance of L1-PCANet and L1-2D2PCANet with PCANet and 2DPCANet as baselines on Yale, AR, extended Yale B, and FERET databases, respectively, which are shown in Fig. 3. SVM29 implementation from the libsvm is used as the classifier with default settings. We repeat some experiments 10 times and calculate the average recognition accuracy and root mean square error (RMSE). In all experiments, we create all PCANet and its different variations instances on MATLAB and other CNNs on Tensorflow.

Fig. 3

Images in three datasets. Top line: Extended Yale B,26 middle line: AR,25 bottom line: FERET.28

JEI_28_2_023016_f003.png

3.1.

Extended Yale B

Extended Yale B consists of 2414 images of 38 individuals captured with different lighting conditions. These pictures are preprocessed to have the same size 48×42 and alignment. The parameters are set as k=5, B=3, L1=L2=4.

In experiment 1, we compare L1-PCANet and L1-2D2PCANet with PCANet and 2DPCANet. We randomly select i=2,3,4,5,6,7 images per individual for training and use the rest for testing. We also create AlexNet30 and GoogleNet11 instances for comparison, which are trained on 1024 images randomly selected from extended Yale B for 20 epochs. The architecture of AlexNet is the same as in Ref. 30 and the architecture of GoogleNet is the same as in Ref. 11. The parameters of two CNNs are set as learning rate=0.0001, batch size=128, drop keep prob.=0.8. The results are shown in Table 1.

Table 1

Experiment 1 on extended Yale B.26

234567
AlexNet85.56±0.53
GoogleNet95.18±0.42
PCANet83.41±5.3184.51±5.7084.42±5.3782.48±7.1884.06±6.2289.56±5.48
2DPCANet97.48±1.0397.34±1.8197.01±1.6496.71±2.4895.16±2.9397.22±2.02
L1-PCANet97.88±0.2297.98±0.2297.88±0.1897.86±0.1797.94±0.1997.90±0.16
L1-2D2PCANet99.67±0.0999.71±0.0799.73±0.0999.73±0.0699.75±0.0699.77±0.07

In experiment 2, to evaluate the robustness of L1-PCANet and L1-2D2PCANet to outliers, we randomly add blockwise noise to the test images to generate test images with outliers. Within each block, the pixel value is randomly set to be 0 or 255. These blocks occupy 10%, 20%, 30%, and 50% of the images and they are added to the random position of the image, respectively, which can be seen in Fig. 4. The results are shown in Table 2.

Fig. 4

Some generalized face images with outliers of extended Yale B:26 (a) 10%; (b) 20%; (c) 30%; and (d) 50%.

JEI_28_2_023016_f004.png

Table 2

Experiment 2 on extended Yale B.26

10%20%30%50%
PCANet92.68±0.4288.51±0.4074.63±0.4844.10±0.76
2DPCANet94.26±0.2588.71±0.5779.54±0.8955.34±0.70
L1-PCANet94.34±0.4091.50±0.5183.58±0.6065.01±0.61
L1-2D2PCANet99.00±0.1598.28±0.1895.73±0.2084.01±0.74

To demonstrate the superiority of the proposed method, we compare L1-PCANet and L1-2D2PCANet with the traditional L1-PCA and L1-2DPCA in experiment 3. We create L1-PCA and L1-2DPCA instances based on Refs. 23 and 24. The parameters of L1-PCA and L1-2DPCA are set as w=100. We randomly select i=2,3,4,5,6,7 images per individual for gallery images and seven images per individual for training. The results are shown in Table 3.

Table 3

Experiment 3 on extended Yale B.26

234567
L1-PCA22.10±1.6932.68±1.6643.23±2.0052.78±1.7059.23±2.1164.49±1.42
L1-2DPCA35.72±2.5043.26±1.9251.72±2.1260.75±1.4265.44±1.8870.60±1.56
L1-PCANet60.83±3.8174.72±2.0783.13±1.8487.90±1.2391.75±1.6294.37±1.04
L1-2D2PCANet76.23±3.4885.20±2.0490.65±1.6593.52±1.1095.62±1.1496.86±0.77

In experiment 4, we examine the impact of the block size B for L1-PCANet and L1-2D2PCANet. The block size changes from 2×2 to 8×8. The results are shown in Fig. 5(a).

Fig. 5

Recognition rate of L1-PCANet and L1-2D2PCANet on extended Yale B and FERET dataset for varying number of block size. (a) Extended Yale B and (b) FERET.

JEI_28_2_023016_f005.png

3.2.

AR

AR face database contains 2600 color images corresponding to 100 people’s faces (50 men and 50 women). It has two session data from two different days and each person in each session has 13 images, including 7 images with only illumination and expression change, 3 images wearing sunglasses, and 3 images wearing scarf. Images show frontal faces with different facial expressions, illumination conditions, and occlusions (sunglasses and scarf). These pictures are preprocessed to 40×30. The parameters are set as k=5, B=4, L1=L2=4, respectively.

In experiment 5, in order to investigate the impact of the choice of training images, we divide the experiment into four groups: (1) In group 1, we randomly select five images with only illumination and expression change from session 1 per individual as training images; (2) in group 2, we randomly select four images with only illumination and expression change and one image wearing sunglasses from session 1 per individual as training images; (3) in group 3, we randomly select four images with only illumination and expression change and one image wearing scarf from session 1 per individual as training images. The remaining images are test samples; and (4) in group 4, we randomly select three images with only illumination and expression change, one image wearing sunglasses and one image wearing scarf from session 1 per individual as training images. The remaining images in session 1 and all images in session 2 are used as test images. We manually select five images from session 1 as the gallery images and keep gallery images of each group the same. The results are shown in Table 4.

Table 4

Experiment 5 on AR.25

No occlusionSunglassScarfSunglass and scarf
PCANet78.63±3.0978.74±4.8479.23±4.4780.40±4.10
2DPCANet82.94±4.3183.85±4.4882.21±2.9783.44±4.27
L1-PCANet87.09±0.5086.73±0.3187.33±0.1286.46±0.22
L1-2D2PCANet89.26±0.3788.59±0.2788.85±0.2888.52±0.19

In order to investigate the impact of the choice of gallery images, experiment 6 is the same as experiment 5 except that the gallery images and the training images are exchanged. We use the remaining images in session 1 and all images in session 2 as test samples. The results are shown in Table 5.

Table 5

Experiment 6 on AR.25

No occlusionSunglassScarfSunglass and scarf
PCANet66.71±0.8769.62±0.6969.59±0.6972.66±0.70
2DPCANet69.24±0.7074.78±0.7072.14±0.9975.51±0.61
L1-PCANet68.56±0.6575.23±0.6072.35±0.7779.34±0.71
L1-2D2PCANet77.08±0.6481.10±0.3778.34±0.6184.17±0.75

3.3.

FERET

This database contains a total of 11338 facial images. They were collected by photographing 994 subjects at various facial angles. We gathered a subset from FERET, which is composed by 1400 images recording of 200 individuals, with each seven images exhibit large variations in facial expression, facial angle, and illumination. This subset is available in our GitHub repository. These pictures are preprocessed to have the same size 40×40 and alignment. The parameters are set as k=5, B=10, L1=L2=4, respectively.

In experiment 7, we divide the experiment into seven groups. The training images of each group consist of 200 images from the subset with different facial angle, expression, and illumination. We use the remaining images in the subset as test images. The results are shown in Table 6.

Table 6

Experiment 7 on FERET.28

1234567AverageRMSE
PCANet75.8376.8376.1768.0073.6769.8379.1174.213.69
2DPCANet73.1776.1776.1773.6778.3373.5074.0075.001.78
L1-PCANet82.8382.1782.0082.5085.0082.5081.8382.690.99
L1-2D2PCANet86.0084.8385.5086.5087.3386.8386.8386.260.81

In experiment 8, we examine the impact of the block size B for L1-PCANet and L1-2D2PCANet. The block size changes from 2×2 to 10×10. The results are shown in Fig. 5(b).

3.4.

Yale

Yale consists of 15 individuals and 11 images for each individual, which shows varying facial expressions and configurations. These pictures are preprocessed to have the same size 32×32. The parameters are set as k=5, B=4, L1=L2=4, respectively.

In experiment 9, we randomly select i=2,3,4,5,6,7 images per individual for training and use the rest for testing. The results are shown in Table 7.

Table 7

Experiment 9 on Yale.26

234567
PCANet86.33±1.8786.75±2.3787.50±1.5887.25±2.1287.25±2.1487.29±2.22
2DPCANet91.33±2.8091.78±1.9490.44±2.5990.67±2.3490.87±2.9091.93±2.13
L1-PCANet91.45±0.8992.00±0.8391.22±0.5491.00±0.4491.89±0.5192.67±0.33
L1-2D2PCANet94.03±0.3295.10±0.4194.95±0.3395.25±0.3295.16±0.4195.66±0.40

3.5.

LFW-a

LFW-a is a version of LFW after alignment with deep funneling. We gathered the individuals, including more than nine images from LFW-a. The parameters are set as k=5, B=3, L1=L2=4, respectively.

In experiment 10, we randomly choose i=3,4,5,6,7 images per individual for gallery images and keep training images of each group the same. The results are shown in Table 8.

Table 8

Experiment 10 on LFW-a.27

34567
PCANet30.07±4.6931.86±5.3534.35±5.9135.71±6.3438.56±6.82
2DPCANet33.00±3.5235.68±3.6439.02±3.7439.92±3.9843.15±4.12
L1-PCANet34.14±0.3936.27±0.2939.08±0.5740.25±0.7744.26±0.81
L1-2D2PCANet39.35±0.2942.20±0.4645.91±0.3446.99±0.4250.12±0.47

4.

Results and Analysis

Tables 1 and 3 show the results of experiments 1 and 3 on extended Yale B, Table 4 shows the result of experiment 5 on AR, Table 6 shows the result of experiment 7 on FERET, Table 7 shows the result on Yale, and Table 8 shows the result on LFW-a.

In these experiments, we changed the training images by random selection. From the results, we can see that the L1-2D2PCANet outperforms L1-PCA, L1-2DPCA, PCANet, 2DPCANet, and L1-PCANet in terms of recognition accuracy and RMSE, because we introduce L1-norm into the network. The two L1-norm-based networks we proposed are far superior to the traditional L2-norm-based networks in terms of RMSE, which means the proposed networks are insensitive to changes in training images. That is, the accuracy of the traditional L2-norm-based networks largely depends on the choice of training images while the L1-norm-based networks we proposed can achieve better and stable accuracy under any training images. A possible explanation of this phenomenon is as follows. In fact, the expression, posture, illumination condition, and occlusion in the images can be regarded as interference or noise in face recognition. This noise degrades L2-norm-based networks much more than it degrades L1-norm-based networks. Therefore, the proposed networks exhibit the superiority when the training images contain some changes in expression, posture, illumination condition, and occlusion.

Table 2 shows the result of experiment 2 on extended Yale B. In this experiment, we randomly add blockwise noise to the test images. From the results, we can see that as the blockwise noise increases from 10% of the image size to 50%, the performance of PCANet, 2DPCANet, and L1-PCANet drops rapidly while L1-2D2PCANet still has good performance. Therefore, it can be considered that L1-2D2PCANet has better robustness against outlier and noise than other three networks.

We also investigate the impact of the choice of gallery images on AR; see Table 4. From the horizontal comparison of Table 5, the more categories the gallery contains, the higher the accuracy is.

Figure 5 shows the result of experiment 4 on extended Yale B and experiment 8 on FERET. When the block is small, the local information cannot be contained perfectly, and it may get more noise when the block is big.

5.

Conclusion

In this paper, we have proposed a deep learning network L1-2D2PCANet, which is a simple but robust method. We use the L1-norm-based 2DPCA14 instead of L2-norm-based 2DPCA15 for the filter learning because of the advantages of L1-norm. It is more robust to outliers than L2-norm. By introducing L1-norm into 2DPCANet,16 we hope the network will inherit such advantages.

To verify the performance of L1-2D2PCANet, we evaluate them on the facial datasets, including AR, extended Yale B, Yale, and FERET, respectively. The results show that L1-2D2PCANet has three distinct advantages over traditional L2-norm-based networks: (1) Statistically, the accuracy of L1-2D2PCANet is higher than that of other networks on all test datasets. (2) L1-2D2PCANet has better robustness to changes in training images compared with the other networks. (3) Compared with the other networks, L1-2D2PCANet has better robustness to noise and outliers. Therefore, L1-2D2PCANet is an efficient and robust network for face recognition.

However, L1-2DPCA brings more computational load to the network, which increases the computational cost of L1-2D2PCANet. Despite this, the computational cost of L1-2D2PCANet is far less than those traditional CNNs, which are based on backpropagation.

In the future work, we will work on the improving of L1-2DPCA algorithm to solve the problem of the computational cost of L1-2D2PCANet.

Acknowledgments

The paper is supported by the National Natural Science Foundation of China (Grant Nos. 61672265 and U1836218), the 111 Project of Ministry of Education of China (Grant No. B12018), UK EPSRC under Grant No. EP/N007743/1, and MURI/EPSRC/dstl under Grant No. EP/R018456/1.

References

1. 

P. A. Devijver and J. Kittler, Pattern Recognition: A Statistical Approach, Prentice Hall International, New Jersey (1982). Google Scholar

2. 

B. D. Ripley, “Pattern recognition and neural networks,” Technometrics, 39 (2), 233 –234 (1999). Google Scholar

3. 

A. K. Jain, R.P.W. Duin and J. Mao, “Statistical pattern recognition: a review,” IEEE Trans. Pattern Anal. Mach. Intell., 22 (1), 4 –37 (2000). https://doi.org/10.1109/34.824819 Google Scholar

4. 

C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics), 049901 Springer-Verlag, New York (2006). Google Scholar

5. 

X.-J. Wu et al., “A new direct LDA (D-LDA) algorithm for feature extraction in face recognition,” in Int. Conf. Pattern Recognit., (2004). https://doi.org/10.1109/ICPR.2004.1333830 Google Scholar

6. 

Y. Yi et al., “Face recognition using spatially smoothed discriminant structure-preserved projections,” J. Electron. Imaging, 23 (2), 023012 (2014). https://doi.org/10.1117/1.JEI.23.2.023012 JEIME5 1017-9909 Google Scholar

7. 

Y. Lecun, Y. Bengio and G. Hinton, “Deep learning,” Nature, 521 (7553), 436 –444 (2015). https://doi.org/10.1038/nature14539 Google Scholar

8. 

D. Silver et al., “Mastering the game of Go without human knowledge,” Nature, 550 (7676), 354 –359 (2017). https://doi.org/10.1038/nature24270 Google Scholar

9. 

S. Lawrence et al., “Face recognition: a convolutional neural-network approach,” IEEE Trans. Neural Networks, 8 (1), 98 –113 (1997). https://doi.org/10.1109/72.554195 ITNNEP 1045-9227 Google Scholar

10. 

N. Kalchbrenner, E. Grefenstette and P. Blunsom, “A convolutional neural network for modelling sentences,” 1 (2014). Google Scholar

11. 

C. Szegedy et al., “Going deeper with convolutions,” in IEEE Comput. Vision and Pattern Recognit., (2015). https://doi.org/10.1109/CVPR.2015.7298594 Google Scholar

12. 

T. H. Chan et al., “PCANet: a simple deep learning baseline for image classification?,” IEEE Trans. Image Process., 24 (12), 5017 –5032 (2015). https://doi.org/10.1109/TIP.2015.2475625 IIPRE4 1057-7149 Google Scholar

13. 

X. J. Wu et al., “A new algorithm for generalized optimal discriminant vectors,” J. Comput. Sci. Technol., 17 (3), 324 –330 (2002). https://doi.org/10.1007/BF02947310 JCTEEM 1000-9000 Google Scholar

14. 

L. Xuelong, P. Yanwei and Y. Yuan, “L1-norm-based 2DPCA,” IEEE Trans. Syst. Man Cybern. Part B Cybern., 40 (4), 1170 –1175 (2010). https://doi.org/10.1109/TSMCB.2009.2035629 Google Scholar

15. 

J. Yang et al., “Two-dimensional PCA: a new approach to appearance-based face representation and recognition,” IEEE Trans. Pattern Anal. Mach. Intell., 26 (1), 131 –137 (2004). https://doi.org/10.1109/TPAMI.2004.1261097 ITPIDJ 0162-8828 Google Scholar

16. 

D. Yu and X. J. Wu, “2DPCANet: a deep leaning network for face recognition,” Multimedia Tools Appl., 77 (10), 12919 –12934 (2018). Google Scholar

17. 

M. Hirose et al., “Principal component analysis for surface reflection components and structure in the facial image and synthesis of the facial image in various ages,” Proc. SPIE, 9398 939809 (2015). https://doi.org/10.1117/12.2076694 PSISDG 0277-786X Google Scholar

18. 

Z. Jia, B. Han and X. Gao, “2DPCANet: dayside aurora classification based on deep learning,” in CCF Chin. Conf. Comput. Vision., 323 –334 (2015). Google Scholar

19. 

Q. R. Zhang, “Two-dimensional parameter principal component analysis for face recognition,” Adv. Mater. Res., 971–973 1838 –1842 (2014). https://doi.org/10.4028/www.scientific.net/AMR.971-973 ADMRBX 0568-0018 Google Scholar

20. 

L. Tian, C. Fan and Y. Ming, “Multiple scales combined principle component analysis deep learning network for face recognition,” J. Electron. Imaging, 25 (2), 023025 (2016). https://doi.org/10.1117/1.JEI.25.2.023025 JEIME5 1017-9909 Google Scholar

21. 

C. Ding, “R1-PCA: rotational invariant L1-norm principal component analysis for robust subspace factorization,” in Int. Conf. Mach. Learn., (2006). Google Scholar

22. 

A. Baccini, P. Besse, A. D. Falguerolles, “A l1-norm PCA and a heuristic approach,” Ordinal and Symbolic Data Analysis, 359 –368 Springer, New York (1996). Google Scholar

23. 

N. Kwak, “Principal component analysis based on L1-norm maximization,” IEEE Trans. Pattern Anal. Mach. Intell., 30 (9), 1672 –1680 (2008). https://doi.org/10.1109/TPAMI.2008.114 ITPIDJ 0162-8828 Google Scholar

24. 

X. Li, Y. Pang and Y. Yuan, “L1-norm-based 2DPCA,” IEEE Trans. Syst. Man Cybern. Part B, 40 (4), 1170 –1175 (2010). https://doi.org/10.1109/TSMCB.2009.2035629 Google Scholar

25. 

A. M. Martinez, “The AR face database,” 24 (1998). Google Scholar

26. 

A. S. Geo et al., “From few to many: illumination cone models for face recognition under variable lighting and pose,” IEEE Trans. Pattern Anal. Mach. Intell., 23 (6), 643 –660 (2001). https://doi.org/10.1109/34.927464 ITPIDJ 0162-8828 Google Scholar

27. 

P. Zhu et al., “Multi-scale patch based collaborative representation for face recognition with margin distribution optimization,” in Eur. Conf. Comput. Vision, 822 –835 (2012). Google Scholar

28. 

P. J. Phillips et al., “The FERET September 1996 database and evaluation procedure,” Lect. Notes Comput. Sci., 1206 395 –402 (1997). https://doi.org/10.1007/BFb0015972 LNCSD9 0302-9743 Google Scholar

29. 

C. A. Burges, “Tutorial on support vector machines for pattern recognition,” Data Mining Knowl. Discovery, 2 121 –167 (1998). https://doi.org/10.1023/A:1009715923555 Google Scholar

30. 

A. Krizhevsky, I. Sutskever and G. Hinton, “ImageNet classification with deep convolutional neural networks,” Adv. Neural Inf. Process. Syst., 25 (2), (2012). Google Scholar

Biography

Yun-Kun Li received his BS degree in microelectronics from the School of Internet of Things Engineering, Jiangnan University, in 2017. He is currently a postgraduate in the Jiangsu Provincial Engineering Laboratory of Pattern Recognition and Computational Intelligence, Jiangnan University. His research interests include pattern recognition and deep learning.

Xiao-Jun Wu received his BS degree in mathematics from Nanjing Normal University, Nanjing, in 1991, and his MS and PhD degrees in pattern recognition and intelligent system from Nanjing University of Science and Technology, Nanjing, in 1996 and 2002, respectively. He has published more than 200 papers in his fields of research. His current research interests include pattern recognition, computer vision, and computational intelligence.

Josef Kittler received his BA, PhD, and DSc degrees from the University of Cambridge, in 1971, 1974, and 1991, respectively. He is currently a professor of machine intelligence with the Centre for Vision, Speech, and Signal Processing, University of Surrey, Guildford, United Kingdom. He has authored the textbook “Pattern Recognition: A Statistical Approach” and over 600 scientific papers. His current research interests include biometrics, video and image database retrieval, medical image analysis, and cognitive vision.

© 2019 SPIE and IS&T 1017-9909/2019/$25.00 © 2019 SPIE and IS&T
Yun-Kun Li, Xiao-Jun Wu, and Josef Kittler "L1-2D2PCANet: a deep learning network for face recognition," Journal of Electronic Imaging 28(2), 023016 (20 March 2019). https://doi.org/10.1117/1.JEI.28.2.023016
Received: 21 December 2018; Accepted: 26 February 2019; Published: 20 March 2019
Lens.org Logo
CITATIONS
Cited by 21 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Facial recognition systems

Principal component analysis

Convolution

Databases

Binary data

Autoregressive models

Lithium


CHORUS Article. This article was made freely available starting 19 March 2020

Back to Top