• 沒有找到結果。

AN IMPROVEMENT TO THE NEAREST NEIGHBOR CLASSIFIER AND FACE RECOGNITION EXPERIMENTS

N/A
N/A
Protected

Academic year: 2021

Share "AN IMPROVEMENT TO THE NEAREST NEIGHBOR CLASSIFIER AND FACE RECOGNITION EXPERIMENTS"

Copied!
6
0
0

加載中.... (立即查看全文)

全文

(1)

Computing, Information and Control ICIC International c 2013 ISSN 1349-4198

Volume 9, Number 2, February 2013 pp. 543–554

AN IMPROVEMENT TO THE NEAREST NEIGHBOR CLASSIFIER AND FACE RECOGNITION EXPERIMENTS

Yong Xu1, Qi Zhu1, Yan Chen1 and Jeng-Shyang Pan2

1Bio-Computing Research Center

2Innovative Information Industry Research Center

Harbin Institute of Technology Shenzhen Graduate School

HIT Campus of Shenzhen University Town, Xili, Shenzhen 518055, P. R. China laterfall2@yahoo.com.cn; ksqiqi@sina.com; jadechenyan@gmail.com; jspan@cc.kuas.edu.tw

Received November 2011; revised March 2012

Abstract. The conventional nearest neighbor classifier (NNC) directly exploits the

dis-tances between the test sample and training samples to perform classification. NNC independently evaluates the distance between the test sample and a training sample. In this paper, we propose to use the classification procedure of sparse representation to im-prove NNC. The proposed method has the following basic idea: the training samples are not uncorrelated and the “distance” between the test sample and a training sample should not be independently calculated and should take into account the relationship between dif-ferent training samples. The proposed method first uses a linear combination of all the training samples to represent the test sample and then exploits modified “distance” to classify the test sample. The method obtains the coefficients of the linear combination by solving a linear system. The method then calculates the distance between the test sample and the result of multiplying each training sample by the corresponding coefficient and assumes that the test sample is from the same class as the training sample that has the minimum distance. The method elaborately modifies NNC and considers the relationship between different training samples, so it is able to produce a higher classification accu-racy. A large number of face recognition experiments on three face image databases show that the maximum difference between the accuracies of the proposed method and NNC is greater than 10%.

Keywords: Face recognition, Nearest neighbor classifier, Sparse representation, Classi-fication

1. Introduction. The image recognition technique [1-3] can be used for a variety of applications such as objection recognition, personal identification and facial expression recognition [4-14]. For many years researchers in the field of image recognition have adopted the following procedures to perform recognition: image capture, feature selec-tion or feature extracselec-tion and classificaselec-tion. Usually these procedures are consecutively implemented and each process is necessary. The nearest neighbor classifier (NNC) is an important classifier. NNC is also one of the oldest and simplest classifiers [15,17]. The nearest neighbors of the sample were used in a number of fields such as image retrieval, image coding, motion control and face recognition [19,22]. NNC first identifies the train-ing sample that is the closest to the test sample and assumes that the test sample is from the same class as this training sample. Since “close” means “similarity”, we can also say that NNC exploits the “similarity” of the test sample and each training sample to perform classification. To determine the nearest neighbors of the sample is the first step of NNC, so it is very crucial. In the past, various ideas and algorithms were proposed for determining the nearest neighbors. For example, D. Omercevic et al. proposed the idea of meaningful nearest neighbors [23]. H. Samet proposed the MaxNearestDist algorithm for

(2)

finding K nearest neighbors [24]. J. Toyama et al. proposed a probably correct approach for greatly reducing the searching time of the nearest neighbor search method [25]. The focus of this approach is to devise the correct set of k-nearest neighbors obtained in high probability. Y.-S. Chen et al. proposed a fast and versatile algorithm to rapidly perform nearest neighbor searches [26]. Besides the methods in these works, many other meth-ods [26-30] have also been developed for computationally efficiently searching the nearest neighbors. We note that most of these methods focus on improving the computation efficiency of the nearest neighbor search.

We note that recently a distinctive image recognition method, the sparse representation (SR) method was proposed [31]. The applications of SR on image recognition such as face recognition have obtained a promising performance [32-34]. However, it seems that it is not very clear why SR can outperform most of previous face recognition methods and different researchers attribute the good performance of SR to different factors. In our opinion one of remarkable advantages of SR is that it uses a novel procedure to classify the test sample. Actually, this method first represents a test sample by using a linear combination of a subset of the training samples. Then it takes the weighted sum of the training sample as an approximation to the test sample and regards the coefficients of the linear combination as the weights. SR calculates the deviation of the test sample from the weighted sum of all the training samples from the same class and classifies the test sample into the class with the minimum deviation. As the weighted sum of a class is also the sum of the contribution in representing the test sample of this class, we refer to this classification procedure as representation-contribution-based classification procedure (RCBCP). We also say that SR consists of a representation procedure and a classification procedure.

The main rationale of RCBCP is that when determining the distances between the test sample and training samples, it takes into account the relationship of different training samples. If some training samples are collinear, RCBCP will use the weights to reflect the collinear nature and will classify the test sample into the class the weighted sum of which provides the best approximation to the test sample. However, the conventional NNC usually separately evaluates the distances between the test sample and training samples, ignoring the similarity and potential relationship between different training samples. The following example is very helpful to illustrate this difference between RCBCP and the conventional NNC: if two training samples have the same minimum Euclidean distances to the test sample, then NNC will be confused in classifying the test sample. However, under the same condition, RCBCP usually obtains two different “distances” and is still able to determine which training sample is closer to the test sample.

In this paper, motivated by RCBCP, we propose to exploit RCBCP to modify NNC. The basic idea is to use a dependent way to determine the “distances” between the test sample and training samples. We first use all of the training samples to represent the test sample, which leads to a linear system. We directly solve this system to obtain the least-squares solution and then exploit the solution and the classification procedure of NNC to classify the test sample. Differing from the conventional NNC, the proposed method calculates the distance between the test sample and the result of multiplying each training sample by the corresponding weight (i.e., a component of the solution vector) and assumes that the test sample is from the same class as the training sample with the minimum distance. The proposed method is very simple and computationally efficient. Our experiments show that the proposed method always achieves a lower rate of classification errors than NNC. This paper also shows that one modification of the proposed method is identical to NNC. This paper not only proposes an improvement to NNC but also has the following contributions: it confirms that RCBCP is very useful for achieving a good face recognition

(3)

performance. Moreover, it also somewhat illustrates that RCBCP is one of the most important advantages of SR.

The rest of the paper is organized as follows. Section 2 describes our method. Section 3 shows the difference between NNC and the proposed method. Section 4 presents the experimental results. Section 5 offers our conclusion.

2. Problem Statement and Preliminaries. Let A1, . . ., An denote all n training

sam-ples in the form of column vectors. We assume that in the original space test sample Y can be approximately represented by a linear combination of all of the training samples. That is,

Y n

i=1βiAi. (1)

βi is the coefficient of the linear combination. We can rewrite (1) as

Y = Aβ, (2)

where β = (β1, . . . , βn)T, A = (A1, . . . , An).

As we know, if ATA is not singular, we can obtain the least squares solution of (2) using β = (ATA)−1ATY . If ATA is nearly singular, we can solve β by β = (ATA + µI)−1ATY ,

where µ is a positive constant and I is the identity. After we obtain β, we calculate Y0 using Y0 = Aβ and refer to it as the result of the linear combination of all of the training samples.

From (1), we know that every training sample makes its own contribution to repre-senting the test sample. The contribution that the ith training sample makes is βiAi.

Moreover, the ability, of representing the test sample, of the ith training sample Ai can

be evaluated by the deviation between βiAi and Y , i.e., ei = ||Y − βiAi||2. Deviation ei

can be also viewed as a measurement of the distance between the test sample and the

ith training sample. We consider that the smaller ei is, the greater ability of representing

the test sample the ith training sample has. We identify the training sample that has the minimum deviation from the test sample and classify Y into the same class as this training sample.

3. Analysis of Our Method. In this section, we show the characteristics and rationale of our method.

3.1. Difference between our method and NNC. Superficially, our method performs somewhat similarly with NNC, because both of them first evaluate the “distances” be-tween the test sample and each training sample and classify the test sample into the same class as the training sample that has the minimum “distance”. However, our method is different from NNC as follows: it does not directly compute the distance between the test sample and each training sample but calculates the distance between the test sample and the result of multiplying each training sample by the corresponding coefficient. Since the sum of all the training samples weighted by the corresponding coefficients well approxi-mates to the original test sample, the result of multiplying each training sample by the corresponding coefficient can be viewed as an optimal approximation, to the original test sample, generated from the training sample. Thus, the deviation between this approxi-mation and the original test sample can be taken as the “distance” between the training sample and test sample. Intuitively, the smaller the “distance”, the more “similar” to the test sample the training sample.

As the weighted sum (i.e., a linear combination) of all the training samples well repre-sents the test sample, we say that all the training samples provide a good representation for the test sample in a competitive way. According to the classification procedure of our

(4)

Table 4. Rates of the classification errors (%) of our method and NNC on the AR database. We took the first 2, 4, 6 and 8 training samples per class and the others as training samples and test samples, respectively.

Database Training samples Our method NNC CBNNC NNL Difference between our method and NNC Difference between our method and CBNNC Difference between our method and NNL AR 2 per class 30.38 39.93 40.03 40.80 9.55 9.65 10.42 AR 4 per class 31.55 42.69 42.69 42.50 11.14 11.14 10.95 AR 6 per class 30.92 38.88 38.92 38.25 7.96 8.0 7.33 AR 8 per class 33.89 41.76 41.76 41.34 7.87 7.87 7.45

training sample. When computing the distance between the test sample and each train-ing sample, the proposed method not only exploits these two samples but also takes into account the relationship between different training samples. As a result, the proposed method can identify the training sample that has the greatest contribution in represent-ing the test sample. A large number of face recognition experiments show that our method always achieves a higher classification accuracy than NNC and the maximum difference between the accuracies of our method and NNC is greater than 10%.

Acknowledgment. This article is partly supported by Program for New Century

Excel-lent TaExcel-lents in University (Nos. NCET-08-0156 and NCET-08-0155), NSFC under Grant nos. 61071179, 61173086, 61020106004, 61001037 and 61173086 as well as the Fundamen-tal Research Funds for the Central Universities (HIT.NSRIF. 2009130).

REFERENCES

[1] D. Zhang, X. Jing and J. Yang, Biometric image discrimination technologies, Idea Group Publishing, Hershey, USA, 2006.

[2] D. Zhang, F. Song, Y. Xu and Z. Liang, Advanced pattern recognition technologies with applications to biometrics, IGI Global, 2008.

[3] A. Khotanzad and Y. C. Hong, Invariant image recognition by Zernikemoments, IEEE Trans. Pattern

Anal. Mach. Intell., vol.12, no.5, pp.489-497, 1990.

[4] J. Yang and C. Liu, Color image discriminant models and algorithms for face recognition, IEEE

Transactions on Neural Networks, vol.19, no.12, pp.2088-2098, 2008.

[5] Y. Xu, L. Yao and D. Zhang, Improving the interest operator for face recognition, Expert System

with Applications, vol.36, no.6, pp.9719-9728, 2009.

[6] C. Liu and J. Yang, ICA color sSpace for pattern recognition, IEEE Transactions on Neural Networks vol.20, no.2, pp.248-257, 2009.

[7] F. Song, D. Zhang, D. Mei and Z. Guo, A multiple maximum scatter difference discriminant criterion for facial feature extraction, IEEE Transactions on Systems, Man, and Cybernetics, Part B, vol.37, no.6, pp.1599-1606, 2007.

[8] M. Wan, Z. Lai, J. Shao and Z. Jin, Two-dimensional local graph embedding discriminant analy-sis (2DLGEDA) with its application to face and palm biometrics, Neurocomputing, vol.73, no.1-3, pp.197-203, 2009.

[9] Q. Gao, L. Zhang, D. Zhang and H. Xu, Independent components extraction from image matrix,

(5)

[10] Y. Xu, D. Zhang and J. Y. Yang, A feature extraction method for use with bimodal biometrics,

Pattern Recognition, vol.43, pp.1106-1115, 2010.

[11] J. Wang, W. Yang, Y. Lin and J. Yang, Two-directional maximum scatter difference discriminant analysis for face recognition, Neurocomputing, vol.72, no.1-3, pp.352-358, 2008.

[12] H. Sellahewa and S. A. Jassim, Image-quality-based adaptive face recognition, IEEE Transactions

on Instrumentation and Measurement, vol.59, no.4, 2010.

[13] J. G. Wang and E. Sung, Facial feature extraction in an infrared image by proxy with a visible face image, IEEE Transactions on Instrumentation and Measurement, vol.56, no.5, 2007.

[14] X. Gao, J. Zhong, D. C. Tao and X.-L. Li, Local face sketch synthesis learning, Neurocomputing, vol.71, no.10-12, pp.1921-1930, 2008.

[15] T. Cover and P. Hart, Nearest neighbor pattern classification, IEEE Transactions in Information

Theory, pp.21-27, 1967.

[16] K. Weinberger, J. Blitzer and L. Saul, Distance metric learning for large margin nearest neighbor classification, Advances in Neural Information Processing Systems, vol.18, pp.1473-1480, 2006. [17] P. Y. Simard, Y. LeCun and J. Decker, Efficient pattern recognition using a new transformation

distance, in Advances in Neural Information Processing Systems 6, S. Hanson, J. Cowan and L. Giles (eds.), Morgan Kaufman, San Mateo, CA, 1993.

[18] H. Li and T. Gao, Improving the B3LYP absorption energies by using the neural network ensemble and k-nearest neighbor approach, ICIC Express Letters, Part B: Applications, vol.2, no.5, pp.1075-1080, 2011.

[19] J. Zou, K. Yuan, S. Jiang and W. Chen, Brain CT image retrieval combined nonnegative tensor factorization with k-nearest neighbor, Journal of Information and Computational Science, vol.6, no.1, pp.273-281, 2009.

[20] J. Nagasue, Y. Konishi, N. Araki, T. Sato and H. Ishigaki, Slope-walking of a biped robot with k nearest neighbor method, ICIC Express Letters, vol.4, no.3(B), pp.893-898, 2010.

[21] Y. Xu and F. Song, Feature extraction based on a linear separability criterion, International Journal

of Innovative Computing, Information and Control, vol.4, no.4, pp.857-865, 2008.

[22] Z.-M. Lu, S.-C. Chu and K.-C. Huang, Equal-average equal-variance equal-norm nearest neighbor codeword search algorithm based on ordered Hadamard transform, International Journal of

Inno-vative Computing, Information and Control, vol.1, no.1, pp.35-41, 2005.

[23] D. Omercevic, O. Drbohlav and A. Leonardis, High-dimensional feature matching: Employing the concept of meaningful nearest neighbors, IEEE the 11th International Conference on Computer

Vision, pp.1-8, 2007.

[24] H. Samet, K-nearest neighbor finding using MaxNearestDist, IEEE Trans. Pattern Anal. Mach.

Intell., vol.30, no.2, pp.243-252, 2008.

[25] J. Toyama, M. Kudo and H. Imai, Probably correct k-nearest neighbor search in high dimensions,

Pattern Recognition, vol.43, pp.1361-1372, 2010.

[26] Y. S. Chen, Y. P. Hung, T. F. Yen and C. S. Fuh, Fast and versatile algorithm for nearest neighbor search based on a lower bound tree, Pattern Recognition, vol.40, pp.360-375, 2007.

[27] S. Berchtold, B. Ertl, D. A. Keim, H.-P. Kriegel and T. Seidl, Fast nearest neighbor search in high-dimensional spaces, Proc. of the 14th IEEE Conference on Data Engineering, pp.23-27, 1998. [28] J. McNames, A fast nearest-neighbor algorithm based on a principal axis search tree, IEEE Trans.

Pattern Anal. Mach. Intell., vol.23, no.9, pp.964-976, 2001.

[29] S. A. Nene and S. K. Nayar, A simple algorithm for nearest neighbor search in high dimensions,

IEEE Trans. Pattern Anal. Mach. Intell., vol.19, pp.989-1003, 1997.

[30] A. Djouadi and E. Boutache, A fast algorithm for the nearest-neighbor classifier, IEEE Trans.

Pattern Anal. Mach. Intell., vol.19, no.3, pp.277-282, 1997.

[31] J. Wright, A. Y. Yang, A. Ganesh et al., Robust face recognition via sparse representation, IEEE

Trans. Pattern Anal. Mach. Intell., vol.31, no.2, pp.210-227, 2009.

[32] J. Wright, Y. Ma, J. Mairal et al., Sparse representation for computer vision and pattern recognition,

Proc. of IEEE, pp.1-8, 2009.

[33] C. X. Ren and D. Q. Dai, Sparse representation by adding noisy duplicates for enhanced face recognition: An elastic net regularization approach, Chinese Conference on Pattern Recognition, Nanjing, China, 2009.

[34] Y. Xu and Q. Zhu, A simple and fast representation-based face recognition method, Neural

Com-puting and Applications, 2012.

[35] http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html. [36] http://cvc.yale.edu/projects/yalefaces/yalefaces.html.

(6)

[37] http://cobweb.ecn.purdue.edu/∼aleix/aleix face DB.html.

[38] J. Yang, D. Zhang, A. F. Frangi and J.-Y. Yang, Two-dimensional PCA: A new approach to appearance-based face representation and recognition, IEEE Trans. Pattern Anal. Mach. Intell., no.1, pp.131-137, 2004.

[39] Q-B Gao, Z. Z. Wang, Center-based nearest neighbor classifier, Pattern Recognition, vol.40, pp.346-349, 2007.

[40] W. Zheng, L. Zhao and C. Zou, Locally nearest neighbor classifiers for pattern classification, Pattern

Recognition, vol.37, pp.1307-1309, 2004.

[41] Y. Xu and Z. Jin, Down-sampling face images and low-resolution face recognition, The 3rd

Inter-national Conference on Innovative Computing, Information and Control, Dalian, China, pp.392,

數據

Table 4. Rates of the classification errors (%) of our method and NNC on the AR database

參考文獻

相關文件

• The start node representing the initial configuration has zero in degree.... The Reachability

了⼀一個方案,用以尋找滿足 Calabi 方程的空 間,這些空間現在通稱為 Calabi-Yau 空間。.

• ‘ content teachers need to support support the learning of those parts of language knowledge that students are missing and that may be preventing them mastering the

 Promote project learning, mathematical modeling, and problem-based learning to strengthen the ability to integrate and apply knowledge and skills, and make. calculated

Wang, Solving pseudomonotone variational inequalities and pseudocon- vex optimization problems using the projection neural network, IEEE Transactions on Neural Networks 17

volume suppressed mass: (TeV) 2 /M P ∼ 10 −4 eV → mm range can be experimentally tested for any number of extra dimensions - Light U(1) gauge bosons: no derivative couplings. =>

Define instead the imaginary.. potential, magnetic field, lattice…) Dirac-BdG Hamiltonian:. with small, and matrix

• Formation of massive primordial stars as origin of objects in the early universe. • Supernova explosions might be visible to the most