Article Preview
TopIntroduction
Automatic Face Recognition (FR) has become one of the most interesting research field that aims essentially to insure security surveillance, telecommunication and human-computer intelligent interaction. Face characteristics of universality, non-intrusively, high social acceptability besides of development of capture devices made the face one of the most widely used biometric technology. Numerous FR techniques have been proposed. They could be classified into two main categories: the geometric based methods and the holistic-based methods. Geometric-based methods extract local features such as eyes, nose, and mouth locations. This category includes local binary patterns (LBP) (Garg & Rajput, 2014), Elastic bunch graph matching (EBGM) (Wiskott, Fellous, Krüger, & Von Der Malsburg, 1997, September), Feature extraction by Gabor filter (Lee, 1996), etc. Holistic-based approaches, the most promising approaches, extract a holistic characteristic of the entire face region. The basic idea of this category consists to transform facial data from high dimensional space to the lower dimensional subspace called feature space in which the researchers apply the classification. This category is itself divided into linear and non-linear approaches. The common linear reduction techniques used are principal component analysis (PCA) (Turk & Pentland, 1991), linear discriminant analysis (LDA) (Belhumeur, Hespanha, & Kriegman, 1997; Lu, Plataniotis, & Venetsanopoulos, 2003), independent component analysis (ICA) (Draper, Baek, Bartlett, & Beveridge, 2003) etc. The non-linear techniques including Exponential discriminant analysis (EDA) (Zhang, Fang, Tang, Shang, & Xu, 2010), Laplacian eigen maps (He, Yan, Hu, Niyogi, & Zhang, 2005; Raducanu & Dornaika, 2010, May), Diffusions maps (Hagen, Smith, Banasuk, Coifman, & Mezie, 2007, December) etc. It should be noted that PCA and LDA are the two popular linear reduction techniques used in face recognition.
Turk and Pentland proposed Eigenface method. They used PCA to reduce the dimension into a face subspace by maximizing the variance across all samples. Then, the images are linearly projected onto this subspace (Belhumeur et al., 1997; Turk & Pentland, 1991). Consequently, LDA method for face classification was developed. LDA based algorithms outperform PCA (Belhumeur et al., 1997). This method focuses on maximizing the Fisher’s criterion, which is the ratio between the between-class scatter matrix and within-class scatter matrix. So unlike PCA technique, LDA maximize the discriminatory features between classes of samples. Nevertheless, the LDA seriously suffers from the so-called “small sample size” (SSS) problem (Chen, Liao, Ko, Lin, & Yu, 2000; Lu et al., 2003) where the number of available training samples is smaller than the dimensionality of the samples (Raudys & Jain, 1991). To overcome the SSS problem, different methods have been proposed. The traditional one is Fisherface (Belhumeur et al., 1997; Zhang & Ruan, 2010), that use PCA as a preprocessing step to reduce the dimension of the within-class scatter matrix by eliminating its null principal components, and then apply the LDA in the retain PCA subspace (Belhumeur et al., 1997). However, the discarded null principal components may contain significant discriminatory information. Researchers (Chen et al., 2000; Yu & Yang, 2001) proposed direct-LDA (D-LDA) method. In which data are performed directly in the original high-dimensional input space. However, D-LDA fails to deliver good performance in case of insufficient training samples. In fact, computing the very small eigenvalues of the within-class scatter matrix give a high variance (Lu, Plataniotis, & Venetsanopoulos, 2005). To overcome this problem, researchers proposed Regularized-LDA (R-LDA) (Lu et al., 2005). They include a regularization parameter () which is used to increase the smaller eigenvalues. This regularization parameter has also the effect of stabilizing the smaller eigenvalues.