Next Article in Journal
Optimizations for Energy Efficiency in Software-Defined Wireless Sensor Networks
Next Article in Special Issue
Deep-Net: A Lightweight CNN-Based Speech Emotion Recognition System Using Deep Frequency Features
Previous Article in Journal
A Robust Multilevel DWT Densely Network for Cardiovascular Disease Classification
Previous Article in Special Issue
Demodulation Method for Loran-C at Low SNR Based on Envelope Correlation–Phase Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Minimum Eigenvector Collaborative Representation Discriminant Projection for Feature Extraction

National Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Submission received: 27 June 2020 / Revised: 17 August 2020 / Accepted: 22 August 2020 / Published: 24 August 2020
(This article belongs to the Special Issue Data, Signal and Image Processing and Applications in Sensors)

Abstract

:
High-dimensional signals, such as image signals and audio signals, usually have a sparse or low-dimensional manifold structure, which can be projected into a low-dimensional subspace to improve the efficiency and effectiveness of data processing. In this paper, we propose a linear dimensionality reduction method—minimum eigenvector collaborative representation discriminant projection—to address high-dimensional feature extraction problems. On the one hand, unlike the existing collaborative representation method, we use the eigenvector corresponding to the smallest non-zero eigenvalue of the sample covariance matrix to reduce the error of collaborative representation. On the other hand, we maintain the collaborative representation relationship of samples in the projection subspace to enhance the discriminability of the extracted features. Also, the between-class scatter of the reconstructed samples is used to improve the robustness of the projection space. The experimental results on the COIL-20 image object database, ORL, and FERET face databases, as well as Isolet database demonstrate the effectiveness of the proposed method, especially in low dimensions and small training sample size.

1. Introduction

High-dimensional data widely exists in real applications, such as image recognition, information retrieval, etc. Particularly, in actual data processing problems, one often encounters the so-called high-dimensiona l small sample size (SSS) problem, in which the number of available samples is smaller than the dimensionality of the sample feature. Besides, high-dimensional data contains a lot of redundant information, and directly processing high-dimensional data will consume a lot of storage and computing resources. Fortunately, some previous research work [1,2,3] has shown that high-dimensional data is likely lying on or close to a low-dimensional submanifold space, which means that high-dimensional data can be projected into a low-dimensional subspace by some dimensionality reduction (DR) methods without losing important information. In the past few decades, numerous DR theories and methods have been proposed, and part of the work can be found in [4,5,6,7,8].
Principal component analysis (PCA) [4] and linear discriminant analysis (LDA) [5] are the most classic and popular DR methods, and PCA belongs to an unsupervised method while LDA is a supervised method. Despite their simplicity and effectiveness, they still suffer from some limitations in practice. For example, PCA is an unsupervised DR method and fails to provide discrimination information for different classes of data. LDA can find at most C 1 meaningful discriminant projection directions because theoretical analysis shows that the rank of the between-class scatter matrix is at most C 1 , where C represents the number of classes. More importantly, they do not use the structural information of the samples, which greatly reduces the discriminativeness of the extracted features, especially when dealing with SSS problems. Most of the DR methods proposed afterward are based on these two methods or their extended version but consider the structure of the samples. Some DR methods which focus on the local structure of the data have been developed to improve the discrimination performance of the low-dimensional space. A partial list of these methods includes the unsupervised ones such as locality preserving projections (LPP) [9], neighborhood preserving embedding (NPE) [10], and locally linear embedding (LLE) [1], and the supervised ones like marginal Fisher analysis (MFA) [11] and constrained discriminant neighborhood embedding (CDEN) [12]. Several other DR methods that utilize both the local and global structure of the data to improve recognition accuracy have been proposed, such as local Fisher discriminant analysis (LFDA) [13], locally linear discriminant embedding (LLDE) [14], and locality preserving discriminant projections (LPDP) [15].
Recently, some representation-based methods have been used for classification, including sparse representation classification (SRC) [16], sparsity preserving projections (SPP) [17], discriminant sparse neighborhood preserving embedding (DSNPE) [18], and discriminative sparsity preserving projections (DSPP) [19]. However, solving a sparse problem requires an iterative method, which is usually time-consuming. Another type of representation-based DR method, which uses L2 regularization and has a closed solution, has attracted wide attention. It has been proved in [20] that the collaborative representation classification (CRC), with higher efficiency, is competitive to SRC in terms of recognition accuracy. Since then, some collaborative representation based have been proposed, such as collaborative representation based projections (CRP) [21], regularized least square-based discriminative projections (RLSDP) [22], a collaborative representation reconstruction based projections (CRRP) [23], collaborative representation based discriminant neighborhood projections (CRDNP) [24] and collaborative preserving Fisher discriminant analysis (CPFDA) [25].
Except for the linear dimensionality reduction methods introduced above, a class of nonlinear dimensionality reduction methods has been proposed to deal with nonlinear dimensionality reduction problems. Most of the nonlinear DR methods directly use kernel trick to expand linear DR method, likes kernel PCA (KPCA) [26], kernel Fisher discriminant (KFD) [27], kernel direct discriminant analysis (KDDA) [28], and kernel collaborative representation-based projection (KCRP) [29]. Some other recent DR methods can be found in [30,31,32].
In this article, we study linear DR methods. Although some of the proposed methods make use of both the local structure and the global structure of the sample to extract features, when the reconstruction error of the sample is large, it is difficult for them to maintain the true structural similarity of the sample. In addition, most of the previous dimensionality reduction methods only consider the structural similarity of similar samples when looking for the projection subspace, while ignoring the structural similarity of different types of samples, which could also be used to improve the discriminability of the extracted features.
In order to extract the features with a strong discriminant, a minimum eigenvector collaborative representation discriminant projection (MECRDP) is proposed in this paper. The main contributions of our work are as follows. First, in the collaborative representation of samples, we not only use the information of the sample space but also consider the information of the sample eigenvector space. Specifically, we use the eigenvector corresponding to the smallest non-zero eigenvalue of the sample covariance matrix to reduce the collaborative representation error of each sample. Then, we maintain the collaborative representation relationship of the samples to improve the discriminability of the extracted features. Also, the between-class scatter of the reconstructed samples is used to improve the robustness of the projection subspace. Last, experimental results on four public databases show that MECRDP outperforms other DR methods in terms of recognition accuracy, especially in low dimensions and small training sample size.
The remainder of this paper is organized as follows. In Section 2, we briefly introduce the research work closely related to our method, including LDA and CRP. Section 3, we propose a minimum eigenvector collaborative representation discriminant projection to improve the discriminability of the projection subspace. The experimental results are presented in Section 4. Finally, the conclusion remarks are given in Section 5.

2. Related Works

For simplicity, suppose a training samples set of C classes is denoted by X = [ x 1 , x 2 , , x n ] m × n , where x i m represents the i th sample, m is the dimension of the sample feature, and n is the number of samples. Besides, suppose the c th class contains n c samples, and c = 1 C n c = n . The method we proposed in this paper has a great relationship with LDA and CRP. In what follows, we briefly review these two methods.

2.1. Linear Discriminant Analysis

The goal of LDA [5] is seeking a projection matrix so that the within-class scatter is minimized, and the between-class scatter is maximized simultaneously. According to the graph embedding [13], the projection matrix of LDA corresponds to the following two optimization problems, respectively,
max P i , j n P T x i P T x j 2 2 W i , j ( b ) ,
min P i , j n P T x i P T x j 2 2 W i , j ( w ) ,
where the weights W i , j ( b ) and W i , j ( w ) are defined as, respectively,
W i , j ( b ) = { 1 / n 1 / n c , if   x i   and   x j belong   to   the   c th   class , 1 / n , otherwise . .
W i , j ( w ) = { 1 / n c ,   if   x i   and   x j belong   to   the   same   class , 0 , otherwise . ,
Using some algebraic transform, we can rewrite (1) and (2) as
max P i , j n P T x i P T x j 2 2 W i , j ( b ) = t r ( P T ( i , j = 1 n ( x i x j ) W i , j ( b ) ( x i x j ) T ) P ) = t r ( P T ( i , j = 1 n ( x i W i , j ( b ) x i T x i W i , j ( b ) x j T x j W i , j ( b ) x i T + x j W i , j ( b ) x j T ) ) P ) = 2 t r ( P T ( X D ( b ) X T X W ( b ) X T ) P ) = 2 t r ( P T ( X ( D ( b ) W ( b ) ) X T ) P ) = 2 t r ( P T ( X L ( b ) X T ) P ) = t r ( P T S b P )
min P i , j n P T x i P T x j 2 2 W i , j ( w ) = t r ( P T ( i , j = 1 n ( x i x j ) W i , j ( w ) ( x i x j ) T ) P ) = t r ( P T ( i , j = 1 n ( x i W i , j ( w ) x i T x i W i , j ( w ) x j T x j W i , j ( w ) x i T + x j W i , j ( w ) x j T ) ) P ) = 2 t r ( P T ( X D ( w ) X T X W ( w ) X T ) P ) = 2 t r ( P T ( X ( D ( w ) W ( w ) ) X T ) P ) = 2 t r ( P T ( X L ( w ) X T ) P ) = t r ( P T S w P )
where t r ( · ) denotes matrix trace, S b = 2 X L ( b ) X T and S w = 2 X L ( w ) X T are the between-class scatter matrix, and the within-class scatter matrix, respectively. L ( b ) = D ( b ) W ( b ) and L ( w ) = D ( w ) W ( w ) are the Laplacian matrices, in which D ( b ) and D ( w ) are diagonal matrices with their diagonal entries are D i , i ( b ) = j = 1 n W i , j ( b ) and D i , i ( w ) = j = 1 n W i , j ( w ) , respectively.
Using (5) and (6), the objective function of LDA can be modeled as
max P t r ( P T S b P ) t r ( P T S w P ) ,

2.2. Collaborative Representation Based Projections

CRP [21] is an unsupervised discriminant projection method based on L2 regularized least squares. The collaborative representation coefficients of the i th sample is gotten by solving the following optimization problem
min r i x i X r i 2 2 + λ r i 2 2 , s . t .    e i T r i = 0 ,
where r i n is the collaborative representation coefficients of the i th sample, λ > 0 is a regularization parameter, and e i = [ 0 1 , , 0 i 1 , 1 , 0 i + 1 , , 0 n ] T . The constraint in (1) means the i th sample is represented on all the samples other than itself. The optimal solution of (1) can be easily achieved by using the Lagrangian multiplier method as
r i = Q ( X x i e i T Q X x i e i T Q e i e i ) ,
where Q = ( X T X + λ I ) 1 and Ι n × n is a identity matrix.
Using the collaborative representation coefficients, the optimal projection matrix P m × d ( d < m ) of CRP is obtained by solving the following two optimization problems, simultaneously
min P i = 1 n P T x i j = 1 n r i , j P T x j 2 2 ,
max P i = 1 n P T x i P T x ¯ 2 2 ,
where r i , j denotes the j th entry of r i , and x ¯ = ( 1 / n ) i = 1 n x i is the mean of the samples.
With some algebraic formulations, (10) and (11) can be, respectively, rewritten as
min P i = 1 n P T x i j = 1 n r i , j P T x j 2 2 = i = 1 n P T X e i P T X r i 2 2 = t r ( P T X ( i = 1 n ( e i r i ) ( e i r i ) T ) X T P ) = t r ( P T X ( I R ) ( I R ) T X T P ) = t r ( P T S L P )
max P i = 1 n P T x i P T x ¯ 2 2 = t r ( P T ( i n ( x i x ¯ ) ( x i x ¯ ) T ) P ) = t r ( P T S T P )
where R = [ r 1 , r 2 , , r n ] is the representation coefficients matrix, S L = ( I R ) ( I R ) T denotes the local scatter matrix and S T = i n ( x i x ¯ ) ( x i x ¯ ) T is the total scatter matrix.
Then, the optimal projection matrix of the CRP is gotten by solving the following optimization problem
min P t r ( P T S L P ) t r ( P T S T P )

3. Minimum Eigenvector Collaborative Representation Discriminant Projection

LDA uses between-class scatter and within-class scatter to improve the discrimination of extracted features, but it ignores the structural relationship of the samples, resulting in a decline in the discrimination of the features in the projection subspace. Although CRP takes into account the structural relationship of the sample, it does not use the class information of the sample, which is not conducive to the improvement of recognition accuracy. In particular, when the number of samples is small, there will be a large representation error, x i X r i 2 2 , which fails to maintain the similarity of the sample structure. Here, we propose a new feature extraction method to alleviate the problems mentioned above.

3.1. Method Proposed

The sample is represented on all samples other than itself; therefore, this may cause a large reconstruction error when the number of samples is small. However, any eigenvector corresponding to a non-zero eigenvalue of the sample covariance matrix contains partial information of each sample. In order to improve the reconstruction error of the sample, but only use very little information of the represented sample, we consider constructing an expanded sample matrix X ˜ with the eigenvector corresponding to the smallest non-zero eigenvalue of the sample covariance matrix. Let x v be the eigenvector corresponding to the smallest non-zero eigenvalue of X X T , then the expanded sample matrix X ˜ is defined as X ˜ = [ X , x v ] . Similar to (8), the collaborative representation coefficients of the i th sample is achieved by
min r ˜ i x i X ˜ r ˜ i 2 2 + λ r ˜ i 2 2 , s . t .    e ˜ i T r ˜ i = 0 ,
where r ˜ i n + 1 is the collaborative representation coefficient and e ˜ i = [ e i T , 0 ] T . Let R ˜ = [ r ˜ 1 , r ˜ 2 , , r ˜ n ] be the collaborative representation coefficient matrix.
Then, the sample x i can be reconstructed as
x ^ i = X ˜ r ˜ i ,
and the reconstructed sample matrix is
X ^ = [ x ^ 1 , x ^ 2 , , x ^ n ] = X [ r ^ 1 , r ^ 2 , , r ^ n ] = X R ˜
In order to maintain the reconstruction similarity of the samples and keep within-class compactness, we modified the optimization problem in (2) as
min P i , j n P T x i P T x ^ j 2 2 W i , j ( w ) ,
Similar to (6), we can simplify (18) as
min P i , j n P T x i P T x ^ j 2 2 W i , j ( w ) = t r ( P T ( i , j n ( x i x ^ j ) W i , j ( w ) ( x i x ^ j ) T ) P ) = t r ( P T ( i , j n ( x i W i , j ( w ) x i T x ^ j W i , j ( w ) x i T x i W i , j ( w ) x ^ j T + x ^ j W i , j ( w ) x ^ j T ) ) P ) = t r ( P T ( X D ( w ) X T X ^ W ( w ) X T X W ( w ) X ^ T + X ^ D ( w ) X ^ T ) P ) = t r ( P T S w P )
where S w = X D ( w ) X T X ^ W ( w ) X T X W ( w ) X ^ T + X ^ D ( w ) X ^ T is the collaborative within-class scatter matrix and W ( w ) , D ( w ) are defined in (4) and (6).
In what follows, we define the collaborative reconstructed between-class scatter matrix as
S ^ b = i , j n ( x ^ i x ^ j ) W i , j ( b ) ( x ^ i x ^ j ) T = i , j n ( x ^ i W i , j ( b ) x ^ i T x ^ j W i , j ( b ) x ^ i T x ^ i W i , j ( b ) x ^ j T + x ^ j W i , j ( b ) x ^ j T ) = X ^ D ( b ) X ^ T X ^ W ( b ) X ^ T X ^ W ( b ) X ^ T + X ^ D ( b ) X ^ T = 2 X ^ ( D ( b ) W ( b ) ) X ^ T = 2 X ^ L ( b ) X ^ T
where L ( b ) has been defined (5).
Based on the Fisher criterion, considering both the between-class scatter matrix S b and the collaborative reconstructed between-class scatter matrix S ^ b , the objective function of the proposed MECRDP is formulated as
min P t r ( P T S w P ) t r ( P T S ˜ b P ) ,
where S ˜ b = α S b + ( 1 α ) S ^ b and α [ 0 , 1 ] is a balance factor between S b and S ^ b . According to the definition of S b , we know that maximizing t r ( P T S b P ) can improve the discrimination of the projection matrix P . However, when the dimension of the subspace exceeds a certain size, as the dimension of the projection subspace increases, the discriminant performance is more affected by noise. Consider that the collaborative representation between samples could better describe the similarity of samples in the structure, thereby reducing the impact of the noise, then we can use S ^ b to improve the robustness of the projection matrix.
The problem (21) is a generalized eigenvalue problem, whose optimal solution could be achieved by the generalized eigenvalue decomposition as follows
S w p i = λ i S ˜ b p i ,
where λ i and p i are the eigenvalue and corresponding eigenvector, respectively. Then the projection matrix P is composed of the eigenvectors corresponding to d —the smallest non-zero eigenvalues—that is P = [ p 1 , p 2 , , p d ] .
Using x v , different from CRP, we not only consider the information in the sample space but also use part of the information in the feature value space, so we can get a smaller representation error, x i X r i 2 2 and keep more collaborative reconstruction information in the feature projection space. In addition, CRP uses the local divergence and global divergence of the sample to obtain the discriminativeness of the extracted features, and in our method, we use the label information of the sample to keep the sample in the projection space with a small within-class scatter and larger between-class scatter, thereby improving the discriminativeness of the extracted features.

3.2. Algorithmic Procedures

For simplicity, we summarize the algorithmic procedures of the proposed MECRDP as follows:
Step 1:
Project the original high-dimensional sample into an intermediate subspace to remove noise and useless information by PCA, and get the projection matrix P V . For simplicity, we still utilize X to represent the training samples after the projection.
Step 2:
Find the eigenvector x v corresponding to the smallest non-zero eigenvalue of X X T , then the expanded sample matrix is X ˜ = [ X , x v ] .
Step 3:
Compute the collaborative representation coefficients R ˜ by (15) and then reconstruct the samples matrix X ^ by X ^ = X R ˜ .
Step 4:
Compute the collaborative within-class scatter matrix S w , the between-class scatter matrix S b and the collaborative reconstructed between-class scatter matrix S ^ b by (19), (5) and (20), respectively.
Step 5:
Perform generalized eigenvalue decomposition by S w p i = λ i S ˜ b p i , and use eigenvectors corresponding to the smallest d eigenvalues to construct the projection matrix P = [ p 1 , p 2 , , p d ] .
Step 6:
For any input sample x n , its low-dimensional projection is y = P T P V T x .
It is obvious that the proposed method has no iterative steps and its optimal projection matrix can be analytically obtained. Without considering the first step, which is the data preprocessing process that most linear DR methods need to perform, we simply analyze the computational complexity of our proposed method. In step 2, it costs about O ( m 3 ) to find the eigenvector corresponding to the smallest non-zero eigenvalue of X X T , step 3 costs about O ( m n 2 + n 3 ) , and step 4 needs about O ( m n 2 + m 2 n ) , finding the optimal projection matrix in step 5 requires O ( n 3 ) , with the total computational complexity being about O ( m 3 + m 2 n + m n 2 + n 3 ) .

4. Experiments

In this section, some experiments are conducted to evaluate the performance of the proposed MECRDP. We compare the recognition performance of these methods on four public databases, including an image object database COIL20 [33], and two face databases ORL [34], FERET [35], and Isolet [36].

4.1. Preprocessing and Parameter Setting

In our experiments, all the images were converted to grayscale images and were resized. In the following experiments, each sample is stacked into a column vector in column order and is normalized. Besides, to avoid the singular problem caused by small sample size problems, we reduce the dimension by remaining 98% data energy by PCA. Each database is randomly divided into a training set and test set, say randomly selecting s -samples from each class to form a training set, and the remaining samples are used as a test set. For simplicity, without additional explanation, the nearest neighbor method is used to classify the test samples. Considering that most methods compared have adjustable parameters, for the sake of fairness, we look for the satisfactory parameters in a larger range for them. For example, we find the neighbor parameter for LLDE from 1 to s 1 , with and empirically set the rescaling coefficient to 1. In MFA, the parameter for the intrinsic graph is empirically set as s 1 and the parameter for the penalty graph is chosen from { 1 C , 3 C , 5 C , 7 C , 9 C } .
Particularly, each experiment was independently repeated 20 times to avoid the bias caused by random selection, and the average results are recorded and reported. All the experiments are implemented in MATLAB R2011b installed on a personal computer (Intel Core i5-4590, 3.30 GHz, 8 Gb RAM).

4.2. Experimental Results and Analysis

4.2.1. Experimental on COIL20 Database

The COIL20 [33] is an object recognition database which has 1440 images from 20 objects with 72 images in each object. We resize each image into 32 × 32 and randomly select 4, 6, 8, 10 images from each class to form the training sample. The parameters λ and α in MECRDP are set to 0.5 and 0.1, respectively.
Table 1 records the maximum average recognition accuracy, the standard derivations, the corresponding dimension, and the average running time of each method on COIL20. Figure 1 shows the average recognition accuracy versus the subspace dimensions on COIL20. From the results in Table 1 and Figure 1, we observe that the proposed MECRDP achieves better recognition accuracy than other methods on COIL20 database. Compared with other methods, the maximum average recognition accuracy of MECRDP is improved about 3%. Table 1 shows the running time of our method is almost the same as other methods, which implies the efficiency of our method. In addition, the optimal recognition result of MECRDP is usually obtained in a lower projection dimension than other methods. The same result could be verified more intuitively in Figure 1. The result in Figure 1 shows that when the dimension of the projection subspace is low, MECRDP is significantly better than other methods. This result means that the features extracted by our method have obvious discriminativeness even in the low-dimensional space. Table 2 shows the maximum average recognition accuracy with different dimension by using nearest neighbor classifier. To evaluate the effectiveness of MECRDP further, we compare it with other dimension reduction methods under another classifier two-layer neural network; the results are shown in Table 3. Table 2 and Table 3 show that the recognition accuracy achieved by the neural network is not as good as that of the nearest neighbor method, but the proposed method always achieves the best results. The result of COIL20 verifies the effectiveness of our method for feature extraction.

4.2.2. Experiment on the ORL Database

The ORL [34] is the face database and consists of 400 images from 40 individuals with 10 images for each individual. These images are varying in lighting, facial expressions, and details. In our experiments, we cropped and resized each image to 32 × 32 , and randomly selected 3, 4, 5, 6 images from each class to form the training set. We set λ and α in MECRDP to 0.05 and 0.9, respectively.
Table 4 reports the maximum average recognition accuracy, the standard derivations, the corresponding dimension, and the average running time of each method on the ORL database. Table 5 and Table 6 list the average recognition rates obtained by the nearest neighbor classifier and neural network classifier for each method, respectively. Figure 2 plots the average recognition accuracy versus the subspace dimensions on ORL. The results show that our method is superior to other methods, especially when the number of training samples is small; the performance improvement of our method is more obvious. For example, when the training samples of each class are 3 and 4, the maximum average recognition accuracy of MECRDP is improved by about 4.1% and 3.1%, respectively. Besides, we note that, for each method, the recognition accuracy is increased as the training sample increases, while that of MECRDP always outperforms other methods. Figure 2 shows that as the projection dimension increases, the performance of some methods, such as LDA, LLDE, and RLSDP, will decrease because they can extract some noise information when the projection dimension is large. However, our method has shown a certain degree of robustness. Comparing Table 5 and Table 6, we find that the recognition accuracy obtained by the nearest neighbor classifier is better than that of the neural network classifier. What’s more, the proposed MECRDP can achieve the best results in almost all the selected feature dimensions. The experimental results on the ORL database verify the advantages of the proposed algorithm.

4.2.3. Experiment on the FERET Database

The FERET [35] is the face database and consists of 13,539 images from 1565 individuals, where we select a subset containing 1400 images from 200 individuals, with seven images for each individual. These images are varying in facial expression, illumination, and pose. We cropped and resized each image to 40 × 40 , and randomly selected 3, 4, 5 images from each class to form the training set. We set λ and α in MECRDP to 0.1 and 0.1, respectively.
Table 7 shows the maximum average recognition accuracy, the standard derivations, the corresponding dimensions, and the average running time of each method on the FERET database. The average recognition rates obtained by the nearest neighbor classifier and neural network classifier for each method are shown in Table 8 and Table 9, respectively. Figure 3 illustrates the average recognition accuracy versus the subspace dimensions on FERET. The results in Table 3 and Figure 3 show that the proposed MECRDP outperforms other methods. It is worth noting that FERET has 200 classes but only seven samples for each class. On the FERET database, the average recognition accuracy is greatly improved by MECRDP. For example, the training samples of each class are 3 and 4, the maximum average recognition accuracy of MECRDP is improved by about 26.2% and 14.1%, respectively. Besides, Figure 3 shows that when the projection dimension increases, the recognition accuracy of MECRDP will decrease but it is still higher than other methods. The results in Table 8 and Table 9 show that the nearest neighbor classifier could achieve higher recognition accuracy than the neural network classifier. Table 8 shows that MECRDP has achieved the highest recognition accuracy in each selected feature dimension. Table 8 shows that MECRDP has achieved the highest recognition accuracy in each selected feature dimension, while Table 9 shows that MECRDP performs better than other feature extraction methods in most cases when the neural network classifiers are used. In general, the experimental results show that the features extracted by MECRDP are more discriminative in most cases.

4.2.4. Experiment on the Isolet Database

The Isolet database [36] was generated as follows. One hundred fifty subjects spoke the name of each letter of the alphabet twice. Hence, there were 52 training examples from each speaker and the size of each sample was 617 × 1. The speakers were grouped into sets of 30 speakers each, and were referred to as isolet1, isolet2, isolet3, isolet4, and isolet5. In our experiment, we compared the performance of each feature extraction method in Isolet1. We set λ and α in MECRDP to 1 and 0.1, respectively.
Table 10 shows the maximum average recognition accuracy, the standard derivations, the corresponding dimensions, and the average running time of each method on the Isolet1 database. Table 11 and Table 12 report the average recognition rates obtained by the nearest neighbor classifier and neural network classifier for each method, respectively. Figure 4 illustrates the average recognition accuracy versus the subspace dimensions in Isolet1. The results in Table 10 show that under different numbers of training samples, the best results are always obtained by MECRDP. Figure 4 shows that as the number of training samples increases, the recognition accuracy of all the feature extraction algorithms is improved. Although the advantages of the proposed MECRDP over other methods will gradually decrease, it can maintain a high recognition accuracy. For example, when the number of training samples for each class is five, the recognition accuracy of MECRDP is improved by at least 3.4% compared to other methods. When the number of training samples for each class is increased to 20, the performance of our method is improved by only 1.2%, and its best recognition accuracy is 94.05%. Table 11 and Table 12 show that MECRDP performs better than other feature extraction methods in most cases. Based on the above experimental results, we believe that the proposed MECRDP is an effective feature extraction method.

4.2.5. Parameter Sensitivity Analysis

Here we analyze the influence of parameters on the proposed algorithm through experiments on the four databases.
Figure 5 and Figure 6 plot the maximum average recognition accuracy of MECRDP versus the parameter λ and α for the four samples of each database, respectively. Figure 4 and Figure 5 show that the parameter selection of MECRDP has a great relationship with the sample database. Within a certain range, such as λ ∊ [0, 1], there is a little change in the recognition accuracy of the four databases. While α has a greater impact on the recognition accuracy for COIL20, ORL, and FERET, particularly for Isolet1, the recognition accuracy of MECRDP was robust for parameter α.

4.2.6. Visualization

In order to compare the distribution of extracted features in low-dimensional space more intuitively, we randomly selected six classes of samples from the ORL database and then projected them into two-dimensional space. Figure 7 shows the projection results of all methods in two dimensions. The results show that most methods have good clustering results except CRP and PCA, which are unsupervised feature extraction methods. This shows that the label information helps to improve the discriminative feature extraction. However, in some classes (class 1 and class 6), some methods, such as CRRP, LDA, MFA, and RLSDP, have not achieved a large between-class distance. Figure 7 shows that the features extracted by MECRDP not only have a small within-class compactness, but also have a larger between-class distance.
In a word, these experimental results imply that using the minimum eigenvector to reduce the sample reconstruction error, and maintaining the collaborative representation relationship between the samples helps to improve the discriminability of the extracted features, especially when the number of training samples is small.

5. Conclusions

A new linear dimensionality reduction method based on minimum eigenvector collaborative representation discriminant projection has been proposed for feature extraction in this paper, which can be viewed as an extended collaborative representation projection method. This method employs the eigenvector corresponding to the smallest non-zero eigenvalue of the sample covariance matrix to reduce the collaborative representation error. Meanwhile, the collaborative representation relationship of the samples is maintained in the projection subspace. In addition, the between-class scatter of the reconstructed samples is used to improve the robustness of the projection subspace. The experimental results on four public databases demonstrated the superiority of the proposed method in terms of recognition accuracy as compared with other commonly used linear DR methods. What’s more, the experiments show that the proposed method is especially suitable for dealing with small sample size problems, and it can also work well when the number of training samples is large. Thus, we believe that MECRDP is a general algorithm for feature extraction.
Note that the proposed MECRDP is a parameterized method, and its performance will inevitably be affected by the choice of the parameters, which also happens to other parameterized methods such as CRP, MFA, and RLSDP. In addition, sometimes, as the feature dimension increases, the performance of the algorithm will decrease. How to utilize other information of the sample, such as spatial distribution information, local structure information, and high-order statistical information to further improve the performance and robustness of the algorithm, is also an interesting direction for future study.

Author Contributions

Conceptualization, Methodology, and Software, H.H.; Data and Formal Analysis, H.H.; Writing—Original Draft Preparation, H.H.; Writing—Review & Editing, H.H. and D.-Z.F.; Supervision, H.H. and D.-Z.F.; Project Administration, D.-Z.F.; Funding Acquisition, D.-Z.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61271293.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Tenenbaum, J.B.; De Silva, V.; Langford, J.C. A global geometric framework for nonlinear dimensionality reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef] [PubMed]
  3. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef] [Green Version]
  4. Turk, M.; Pentland, A. Eigenfaces for recognition. J. Cogn. Neurosci. 1991, 3, 71–86. [Google Scholar] [CrossRef] [PubMed]
  5. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D.J. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 711–720. [Google Scholar] [CrossRef] [Green Version]
  6. Hyvärinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar] [CrossRef] [Green Version]
  7. Li, X.; Shu, L. Kernel Based Nonlinear Dimensionality Reduction and Classification for Genomic Microarray. Sensors 2008, 8, 4186–4200. [Google Scholar] [CrossRef] [Green Version]
  8. Rosdi, B.A.; Shing, C.W.; Suandi, S.A. Finger vein recognition using local line binary pattern. Sensors 2011, 11, 11357–11371. [Google Scholar] [CrossRef] [Green Version]
  9. He, X.; Yan, S.; Hu, Y.; Niyogi, P.; Zhang, H.-J. Face recognition using Laplacianfaces. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 328–340. [Google Scholar]
  10. He, X.; Cai, D.; Yan, S.; Zhang, H.-J. Neighborhood preserving embedding. In Proceedings of the 10th IEEE International Conference on Computer Vision (ICCV 2005), Beijing, China, 17–20 October 2005; pp. 1208–1213. [Google Scholar]
  11. Yan, S.; Xu, D.; Zhang, B.; Zhang, H.-J.; Yang, Q.; Lin, S. Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 40–51. [Google Scholar] [CrossRef] [Green Version]
  12. Li, B.; Lei, L.; Zhang, X.-P. Constrained discriminant neighborhood embedding for high dimensional data feature extraction. Neurocomputing 2016, 173, 137–144. [Google Scholar] [CrossRef]
  13. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  14. Li, B.; Zheng, C.-H.; Huang, D.-S. Locally linear discriminant embedding: An efficient method for face recognition. Pattern Recognit. 2008, 41, 3813–3821. [Google Scholar] [CrossRef]
  15. Gui, J.; Jia, W.; Zhu, L.; Wang, S.-L.; Huang, D.-S. Locality preserving discriminant projections for face and palmprint recognition. Neurocomputing 2010, 73, 2696–2707. [Google Scholar] [CrossRef]
  16. Zhang, L.; Chen, S.; Qiao, L. Graph optimization for dimensionality reduction with sparsity constraints. Pattern Recognit. 2012, 45, 1205–1210. [Google Scholar] [CrossRef]
  17. Qiao, L.; Chen, S.; Tan, X. Sparsity preserving projections with applications to face recognition. Pattern Recognit. 2010, 43, 331–341. [Google Scholar] [CrossRef] [Green Version]
  18. Gui, J.; Sun, Z.; Jia, W.; Hu, R.; Lei, Y.; Ji, S. Discriminant sparse neighborhood preserving embedding for face recognition. Pattern Recognit. 2012, 45, 2884–2893. [Google Scholar] [CrossRef]
  19. Gao, Q.; Huang, Y.; Zhang, H.; Hong, X.K.; Wang, L.Y. Discriminative sparsity preserving projections for image recognition. Pattern Recognit. 2015, 48, 2543–2553. [Google Scholar] [CrossRef]
  20. Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the IEEE International Conference on Computer Vision Workshops, ICCV 2011 Workshops, Barcelona, Spain, 6–13 November 2011; pp. 471–478. [Google Scholar]
  21. Yang, W.; Wang, Z.; Sun, C. A collaborative representation based projections method for feature extraction. Pattern Recognit. 2015, 48, 20–27. [Google Scholar] [CrossRef]
  22. Yang, W.; Sun, C.; Zheng, W. A regularized least square based discriminative projections for feature extraction. Neurocomputing 2016, 175, 198–205. [Google Scholar] [CrossRef]
  23. Yin, J.; Wei, L.; Song, M.; Zeng, W. Optimized projection for collaborative representation based classification and its applications to face recognition. Pattern Recognit. Lett. 2016, 73, 83–90. [Google Scholar] [CrossRef]
  24. Wang, G.; Shi, N. Collaborative representation-based discriminant neighborhood projections for face recognition. Neural Comput. Appl. 2020, 32, 5815–5832. [Google Scholar] [CrossRef]
  25. Yuan, M.D.; Feng, D.Z.; Shi, Y.; Liu, W.J. Dimensionality reduction by collaborative preserving Fisher discriminant analysis. Neurocomputing 2019, 356, 228–243. [Google Scholar] [CrossRef]
  26. Kim, K.I.; Jung, K.; Kim, H.J. Face recognition using kernel principal component analysis. IEEE Signal. Process. Lett. 2002, 9, 40–42. [Google Scholar]
  27. Mika, S.; Rätsch, G.; Weston, J.; Schölkopf, B.; Müller, K.-R. Fisher discriminant analysis with kernels. In Proceedings of the IEEE Signal Processing Society Workshop on Neural Network for Signal Processing, Madison, WI, USA, 23–25 August 1998; pp. 41–48. [Google Scholar]
  28. Lu, J.; Plataniotis, K.N.; Venetsanopoulos, A.N. Face recognition using kernel direct discriminant analysis algorithms. IEEE Trans. Neural Netw. 2003, 14, 117–126. [Google Scholar]
  29. Jun, Y.; Lai, Z.; Yan, H. Dimension Reduction using Kernel Collaborative Representation based Projection. AEU Int. J. Electron. Commun. 2017, 81, 23–30. [Google Scholar]
  30. Hoppenstedt, B.; Reichert, M.; Kammerer, K.; Probst, T.; Schlee, W.; Spiliopoulou, M.; Pryss, R. Dimensionality Reduction and Subspace Clustering in Mixed Reality for Condition Monitoring of High-Dimensional Production Data. Sensors 2019, 19, 3903. [Google Scholar] [CrossRef] [Green Version]
  31. Yu, W.; Miao, Z.; Yi, S. Learning a local manifold representation based on improved neighborhood rough set and LLE for hyperspectral dimensionality reduction. Signal. Process. 2019, 164, 20–29. [Google Scholar] [CrossRef]
  32. Xu, X.; Yang, Y.; Deng, C.; Nie, F. Adaptive graph weighting for multi-view dimensionality reduction. Signal. Process. 2019, 165, 186–196. [Google Scholar] [CrossRef]
  33. Nene, S.A.; Nayar, S.K.; Murase, H. Technical Report CUCS-005-96, Columbia Object Image Library (COIL-20); Springer: New York, NY, USA, 1996. [Google Scholar]
  34. Samaria, F.S.; Harter, A.C. Parameterisation of a stochastic model for human face identification. In Proceedings of the Second IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 1 January 1994; pp. 138–142. [Google Scholar]
  35. Phillips, P.J.; Wechsler, H.; Huang, J.; Rauss, P.J. The FERET database and evaluation procedure for face-recognition algorithms. Image Vis. Comput. 1998, 16, 295–306. [Google Scholar] [CrossRef]
  36. Dua, D.; Graff, C. UCI Machine Learning Repository. Irvine, CA: University of California. School of Information and Computer Science. 2019. Available online: http://archive.ics.uci.edu/ml (accessed on 16 August 2020).
Figure 1. The average recognition accuracy vs. the subspace dimensions of each method in the COIL20 database.
Figure 1. The average recognition accuracy vs. the subspace dimensions of each method in the COIL20 database.
Sensors 20 04778 g001
Figure 2. The average recognition accuracy vs. the subspace dimensions of each method in the ORL database.
Figure 2. The average recognition accuracy vs. the subspace dimensions of each method in the ORL database.
Sensors 20 04778 g002
Figure 3. The average recognition accuracy vs. the subspace dimensions of each method in the FERET database.
Figure 3. The average recognition accuracy vs. the subspace dimensions of each method in the FERET database.
Sensors 20 04778 g003
Figure 4. The average recognition accuracy vs. the subspace dimensions of each method in the Isolet1 database.
Figure 4. The average recognition accuracy vs. the subspace dimensions of each method in the Isolet1 database.
Sensors 20 04778 g004aSensors 20 04778 g004b
Figure 5. The maximal average recognition accuracy of MECRDP vs. the regularization parameter λ in the four databases.
Figure 5. The maximal average recognition accuracy of MECRDP vs. the regularization parameter λ in the four databases.
Sensors 20 04778 g005
Figure 6. The maximal average recognition accuracy of MECRDP vs. the balance factor α in the four databases.
Figure 6. The maximal average recognition accuracy of MECRDP vs. the balance factor α in the four databases.
Sensors 20 04778 g006
Figure 7. The visualization of six individuals in the ORL database. (a) CRP, (b) CRRP, (c) LDA, (d) LLDE, (e) MFA, (f) RLSDP, (g) PCA, (h) MECRDP.
Figure 7. The visualization of six individuals in the ORL database. (a) CRP, (b) CRRP, (c) LDA, (d) LLDE, (e) MFA, (f) RLSDP, (g) PCA, (h) MECRDP.
Sensors 20 04778 g007
Table 1. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the COIL20 database.
Table 1. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the COIL20 database.
Methods4-Samples6-Samples8-Samples10-Samples
CRP81.51 ± 2.08 (16, 0.008)85.35 ± 2.12 (14, 0.011)89.90 ± 1.07 (16, 0.017)91.84 ± 1.39 (17, 0.025)
CRRP71.70 ± 2.90 (18, 0.007)77.75 ± 2.59 (18, 0.016)82.12 ± 2.04 (19, 0.026)85.47 ± 1.54 (19, 0.042)
LDA78.36 ± 2.93 (19, 0.008)84.70 ± 2.28 (22, 0.008)88.73 ± 1.74 (22, 0.016)91.30 ± 1.55 (19, 0.017)
LLDE78.22 ± 3.16 (20, 0.009)84.27 ± 2.39 (19, 0.015)88.78 ± 1.73 (19, 0.019)91.38 ± 1.59 (18, 0.029)
MFA78.84 ± 2.86 (26, 0.005)84.65 ± 2.13 (32, 0.010)88.95 ± 1.80 (21, 0.017)91.33 ± 1.35 (18, 0.023)
RLSDP77.25 ± 3.32 (19, 0.005)82.91 ± 2.50 (18, 0.010)87.30 ± 1.85 (20, 0.013)89.83 ± 1.73 (16, 0.020)
PCA80.84 ± 1.75 (18, 0.004)85.67 ± 1.68 (14, 0.006)89.10 ± 1.38 (23, 0.012)91.14 ± 1.25 (18, 0.016)
MECRDP84.89 ± 2.26 (11, 0.007)89.60 ± 2.00 (11, 0.010)92.79 ± 1.36 (13, 0.018)94.50 ± 1.07 (13, 0.026)
Table 2. The maximum average recognition accuracy (%) in the COIL20 database with different dimension by using nearest neighbor classifier.
Table 2. The maximum average recognition accuracy (%) in the COIL20 database with different dimension by using nearest neighbor classifier.
Dimension510152025303540
CRP84.3190.1391.8291.7591.5291.3891.0590.72
CRRP62.0276.2582.7585.2984.4884.1583.8383.41
LDA85.8390.4790.9291.2791.0790.5989.8089.22
LLDE86.2790.8091.1491.1990.4589.6589.0788.61
MFA84.7390.1091.1391.2891.0690.9690.8890.73
RLSDP81.1488.8989.7689.7789.2389.0088.3887.97
PCA83.9089.6690.9891.0290.9990.7690.6190.46
MECRDP90.8094.4094.2793.7193.4293.2093.6992.20
Table 3. The maximum average recognition accuracy (%) on the COIL20 database with different dimension by using neural network classifier.
Table 3. The maximum average recognition accuracy (%) on the COIL20 database with different dimension by using neural network classifier.
Dimension510152025303540
CRP77.2187.1689.6389.1288.1587.4886.3786.05
CRRP60.7674.6079.7180.6780.5779.9879.2778.78
LDA84.4687.1985.1983.1082.7382.6581.7181.84
LLDE85.4887.7585.1383.5182.6282.6682.5780.67
MFA83.6287.5286.1983.4082.9882.3082.0081.44
RLSDP79.9385.7984.6381.5881.8581.8881.5781.02
PCA77.7186.7089.8589.3888.2587.6686.5885.83
MECRDP88.7791.6791.3690.8589.7889.1488.0387.49
Table 4. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the ORL database.
Table 4. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the ORL database.
Methods3-Samples4-Samples5-Samples6-Samples
CRP79.55 ± 3.12 (42, 0.019)84.48 ± 2.32 (52, 0.024)88.08 ± 2.17 (62, 0.044)90.56 ± 3.06 (72, 0.058)
CRRP81.02 ± 2.79 (40, 0.023)87.46 ± 2.25 (40, 0.034)90.58 ± 1.85 (40, 0.074)92.75 ± 1.71 (38, 0.085)
LDA84.20 ± 2.87 (38, 0.013)88.75 ± 4.18 (42, 0.018)92.95 ± 1.85 (42, 0.030)95.25 ± 1.68 (44, 0.035)
LLDE78.66 ± 3.26 (34, 0.016)89.23 ± 2.00 (40, 0.025)92.80 ± 1.57 (40, 0.045)94.97 ± 1.62 (44, 0.054)
MFA84.07 ± 3.19 (58, 0.018)89.50 ± 1.43 (48, 0.027)92.90 ± 1.55 (42, 0.034)95.03 ± 1.17 (48, 0.045)
RLSDP82.86 ± 2.86 (40, 0.012)88.94 ± 2.34 (40, 0.018)92.78 ± 1.57 (40, 0.031)94.62 ± 1.45 (38, 0.038)
PCA75.07 ± 2.95 (80, 0.008)81.44 ± 2.63 (74, 0.012)85.00 ± 2.81 (68, 0.027)88.22 ± 2.54 (78, 0.024)
MECRDP87.92 ± 2.26 (38, 0.014)93.43 ± 1.62 (40, 0.023)95.28 ± 1.70 (40, 0.037)96.66 ± 1.42 (38, 0.054)
Table 5. The maximum average recognition accuracy (%) in the ORL database with different dimensions by using the nearest neighbor classifier.
Table 5. The maximum average recognition accuracy (%) in the ORL database with different dimensions by using the nearest neighbor classifier.
Dimension1020304050607080
CRP81.7887.5388.2588.8789.1990.1690.2890.25
CRRP72.1686.0390.5692.6992.2892.0091.9791.84
LDA88.4193.1293.7595.0395.0094.5093.8492.81
LLDE88.2893.6393.8494.9494.3493.9793.7593.25
MFA89.0093.5094.0694.7594.9794.8194.8194.91
RLSDP88.3492.5693.5094.4194.1294.0992.8192.31
PCA80.9186.1987.2287.6987.8788.0088.1988.19
MECRDP94.1696.2596.2596.6296.5996.0696.0095.91
Table 6. The maximum average recognition accuracy (%) in the ORL database with different dimensions by using the neural network.
Table 6. The maximum average recognition accuracy (%) in the ORL database with different dimensions by using the neural network.
Dimension1020304050607080
CRP81.4788.0986.7284.8883.1380.5676.6975.94
CRRP71.0081.2880.8778.1377.5376.1972.4773.56
LDA85.7287.5386.0678.2877.6976.3875.2874.69
LLDE85.9788.7884.3178.0978.1377.4474.9473.44
MFA86.9788.0684.6678.9776.1273.1970.8872.34
RLSDP85.6287.3883.5379.0378.3176.0974.9774.31
PCA80.5988.4486.9484.3882.1281.0979.7578.22
MECRDP92.0390.5088.3986.1584.1182.2879.6378.47
Table 7. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the FERET database.
Table 7. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimension, and the average running time (seconds) in parentheses of each method in the FERET database.
Methods3-Samples4-Samples5-Samples
CRP31.45 ± 1.40 (80, 0.561)34.99 ± 1.63 (80, 1.189)37.85 ± 2.67 (80, 2.005)
CRRP48.08 ± 1.61 (34, 1.362)64.32 ± 1.69 (32, 2.479)72.91 ± 1.31 (24, 3.924)
LDA52.03 ± 1.82 (34, 0.443)68.64 ± 1.92 (32, 0.809)77.30 ± 1.82 (26, 1.314)
LLDE52.32 ± 1.86 (34, 0.551)68.89 ± 1.61 (26, 1.029)77.84 ± 1.89 (24, 1.630)
MFA52.28 ± 2.04 (36, 0.454)68.95 ± 1.99 (22, 0.864)78.19 ± 1.57 (24, 1.387)
RLSDP48.57 ± 1.71 (36, 0.480)66.35 ± 1.54 (26, 0.873)76.01 ± 1.27 (24, 1.374)
PCA26.52 ± 0.99 (80, 0.245)30.15 ± 1.11 (80, 0.460)34.45 ± 2.03 (80, 0.806)
MECRDP78.13 ± 1.71 (18, 0.563)83.66 ± 1.12 (20, 1.038)87.00 ± 1.26 (24, 1.727)
Table 8. The maximum average recognition accuracy (%) in the FERET database with different dimensions by using the nearest neighbor classifier.
Table 8. The maximum average recognition accuracy (%) in the FERET database with different dimensions by using the nearest neighbor classifier.
Dimension1020304050607080
CRP23.1828.3630.0132.5634.4035.7036.9937.85
CRRP66.7574.4972.5871.6670.6669.5568.6567.42
LDA70.6576.8976.9976.0575.1073.9372.8571.42
LLDE71.0977.5777.8076.5675.3674.6973.5072.55
MFA71.1077.7177.7477.0575.4474.1373.1471.73
RLSDP69.6875.4375.5374.5173.3072.1471.3970.28
PCA25.3529.2130.1632.5633.3933.7634.2134.45
MECRDP80.5586.8986.1184.5582.4279.8177.5675.65
Table 9. The maximum average recognition accuracy (%) in the FERET database with different dimensions by using the neural network.
Table 9. The maximum average recognition accuracy (%) in the FERET database with different dimensions by using the neural network.
Dimension1020304050607080
CRP17.0635.7842.6346.0947.2645.7343.1641.14
CRRP51.8358.0052.8847.1742.0437.9134.2831.39
LDA55.0361.8656.4550.1645.0040.0536.5033.14
LLDE54.8861.8656.1150.1545.2240.3735.7432.75
MFA56.8363.6957.9851.3745.8140.2135.3332.76
RLSDP53.8559.8454.4848.8843.1439.6835.4532.90
PCA17.6535.9944.1847.0947.5146.8045.0543.19
MECRDP68.9178.8872.3065.0657.8550.2644.9042.92
Table 10. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimensions, and the average running time (seconds) in parentheses of each method in the Isolet1 database.
Table 10. The maximum average recognition accuracy (%) ± the standard derivations (%), the corresponding dimensions, and the average running time (seconds) in parentheses of each method in the Isolet1 database.
Methods5-Samples10-Samples15-Samples20-Samples
CRP71.59 ± 1.24 (35, 0.015)76.11 ± 1.39 (39, 0.054)78.80 ± 0.78 (40, 0.125)80.72 ± 1.01 (40, 0.209)
CRRP70.47 ± 2.25 (24, 0.025)79.59 ± 1.21 (24, 0.073)83.79 ± 1.17 (25, 0.223)86.25 ± 1.10 (24, 0.412)
LDA82.59 ± 1.87 (25, 0.012)89.12 ± 1.08 (25, 0.030)91.55 ± 0.76 (25, 0.064)92.80 ± 0.68 (24, 0.118)
LLDE82.48 ± 1.89 (26, 0.016)88.89 ± 0.94 (26, 0.055)91.65 ± 0.75 (26, 0.118)92.86 ± 0.59 (26, 0.217)
MFA82.94 ± 1.72 (29, 0.013)88.93 ± 0.98 (25, 0.043)91.59 ± 0.87 (25, 0.099)92.71 ± 0.82 (24, 0.169)
RLSDP82.31 ± 1.72 (25, 0.011)88.71 ± 0.97 (26, 0.034)91.42 ± 0.53 (25, 0.077)92.80 ± 0.67 (25, 0.144)
PCA71.29 ± 1.57 (40, 0.007)75.86 ± 1.45 (40, 0.022)79.09 ± 0.87 (39, 0.049)80.91 ± 1.13 (40, 0.089)
MECRDP86.40 ± 1.61 (25, 0.013)91.36 ± 1.05 (26, 0.046)93.23 ± 0.77 (25, 0.103)94.05 ± 0.68 (25, 0.199)
Table 11. The maximum average recognition accuracy (%) in the Isolet1 database with different dimensions by using the nearest neighbor classifier.
Table 11. The maximum average recognition accuracy (%) in the Isolet1 database with different dimensions by using the nearest neighbor classifier.
Dimension510152025303540
CRP59.2567.7473.9276.9779.1580.0080.3480.72
CRRP48.0466.5175.9683.8186.2386.0285.7585.59
LDA75.0384.0489.8891.8392.7992.7392.6492.36
LLDE73.0182.8889.4091.5892.8391.6990.5889.92
MFA72.2085.1290.2892.0392.6892.3792.2492.33
RLSDP75.9484.2189.9191.7592.8092.3792.1491.63
PCA58.5167.8374.0877.2579.3280.0480.2380.91
MECRDP74.7286.6291.8293.2694.0593.8293.7393.57
Table 12. The maximum average recognition accuracy (%) in the Isolet1 database with different dimensions by using the neural network classifier.
Table 12. The maximum average recognition accuracy (%) in the Isolet1 database with different dimensions by using the neural network classifier.
Dimension510152025303540
CRP63.2974.5783.8185.7487.1687.4187.5587.13
CRRP50.4066.6675.3580.5981.9281.4080.4479.51
LDA73.7981.8685.6586.4085.5084.8984.0783.75
LLDE70.6680.7585.8686.7086.1485.3985.0884.30
MFA71.5682.9786.5887.6386.3884.9584.0483.00
RLSDP74.2481.7885.8686.2985.9085.1084.1783.53
PCA63.6575.1784.3486.2987.0887.8187.7587.83
MECRDP73.4985.0889.1788.9489.5489.1989.2389.06

Share and Cite

MDPI and ACS Style

Hu, H.; Feng, D.-Z. Minimum Eigenvector Collaborative Representation Discriminant Projection for Feature Extraction. Sensors 2020, 20, 4778. https://0-doi-org.brum.beds.ac.uk/10.3390/s20174778

AMA Style

Hu H, Feng D-Z. Minimum Eigenvector Collaborative Representation Discriminant Projection for Feature Extraction. Sensors. 2020; 20(17):4778. https://0-doi-org.brum.beds.ac.uk/10.3390/s20174778

Chicago/Turabian Style

Hu, Haoshuang, and Da-Zheng Feng. 2020. "Minimum Eigenvector Collaborative Representation Discriminant Projection for Feature Extraction" Sensors 20, no. 17: 4778. https://0-doi-org.brum.beds.ac.uk/10.3390/s20174778

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop