Next Article in Journal
New Indirect Tire Pressure Monitoring System Enabled by Adaptive Extended Kalman Filtering of Vehicle Suspension Systems
Next Article in Special Issue
Automatic Face Mask Detection System in Public Transportation in Smart Cities Using IoT and Deep Learning
Previous Article in Journal
Autonomous Haulage Systems in the Mining Industry: Cybersecurity, Communication and Safety Issues and Challenges
Previous Article in Special Issue
Novel BSSSO-Based Deep Convolutional Neural Network for Face Recognition with Multiple Disturbing Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis of Random Local Descriptors in Face Recognition

1
Instituto Politécnico Nacional (IPN), Ciudad de México 07738, Mexico
2
Department of Micro and Nanotechnology, Institute of Applied Sciences and Technology (ICAT), National Autonomous University of Mexico (UNAM), Mexico City 04510, Mexico
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(11), 1358; https://doi.org/10.3390/electronics10111358
Submission received: 8 May 2021 / Revised: 3 June 2021 / Accepted: 4 June 2021 / Published: 7 June 2021
(This article belongs to the Special Issue Face Recognition Using Machine Learning)

Abstract

:
This article describes and analyzes the new feature extraction technique, Random Local Descriptor (RLD), that is used for the Permutation Coding Neural Classifier (PCNC), and compares it with Local Binary Pattern (LBP-based) feature extraction. The paper presents a model of face feature detection using local descriptors, and describes an improvement on the PCNC for the recognition of plane rotated and small displaced face images, as applied to three databases, i.e., ORL, FRAV3D and FEI. All databases are described along with the recognition results that were obtained. We also include a comparison of our classifier with the Support Vector Machine (SVM) and Iterative Closest Point (ICP). The ORL database was selected to compare our RLDs with LBP-based algorithms. The PCNC with the RLDs demonstrated the best recognition rate, i.e., 97.49%, in comparison with 90.49% for LBPs. For the FEI image database, we obtained the best recognition rate, i.e., 93.57%, in comparison with 66.74% for LBPs. Using the RLDs and rotating the original images for FRAV3D, we improved the recognition rate by decreasing by approximately twice the number of errors. In addition, we analyzed the influence of different RLD parameters on the quality of facial recognition.

1. Introduction

Face recognition offers the advantage of being a passive identification and verification method that does not require explicit action or participation by the individual in order to be recognized. This characteristic makes this technique ideal for security and surveillance purposes. The acquisition methods for face images can be easily performed with inexpensive standard cameras at a long distance. However, unconstrained situations make it difficult to develop robust systems that are invariant to illumination, size, pose and location. Images in the real world are affected by expressions, poses, occlusions, and illumination; the differences among various face images of the same person could be even larger than those from images of a different person altogether. Therefore, extracting robust and discriminative features that make it possible to distinguish among different people is a critical and difficult problem in face recognition [1].
Automatic face recognition could be used in different security systems, such as security for buildings, offices and banks. Different approaches have been investigated and proposed for solving this task [2,3]. Face analysis systems are often built into mobile phones. Although the memory capabilities of mobile phones are limited, experiments show encouraging face detection performance.
Face recognition is a classical recognition process that involves two critical problems—feature representation and classifier construction [1]. It has been demonstratedthat different methods of feature extraction can be combined with different methods of classification in a handwritten digit recognition task.
The authors [4] consider that the design of effective features is a fundamental issue in computer vision. It is commonly accepted that designing effective features has an important tradeoff with discriminativeness and robustness.
The techniques developed thus far for face representation can be roughly classified into two main categories: holistic- and local-based techniques. Holistic approaches are based on the global use of the whole face region. Local-based techniques locate a number of features from a face and then classify them by combining and comparing them with corresponding local statistics. It has been proven that component-based (local feature-based) face recognition methods perform better than global methods (holistic-based).
Different face recognition systems, such as SpereFace, Arcface or Cosface, based on neural networks and deep learning, have been developed in recent years [5,6,7,8,9].
Here, we focus on aspects of feature extraction for facial recognition (classification) using local feature-based methods.
Recently, there has been substantial interest in object and view matching using local invariant features or local feature descriptors [1,4,10,11]. The methods that use these descriptors can be divided into two classes: sparse descriptors, which first detects the interest points in a given image and then samples a local patch and describes its invariant features [10]; and dense descriptors, which extract local features pixel by pixel.
As examples of sparse descriptors, we can mention scale-invariant feature transform (SIFT) and rotation-invariant feature transform (RIFT) [10,12]. A typical image of size 500 × 500 pixels gives rise to about 2000 stable features.
Examples of dense descriptors are Local Binary Pattern Descriptor (LBP) [13,14] and Weber Local Descriptor (WLD) [10]. LBP is one of the most powerful descriptors to represent local structures [13,14]. An example of a basic LBP operator is presented in Figure 1.
We have a window of 3 × 3 pixels in which every pixel has a brightness value, which is compared with the brightness of the central point. The value of the central point is a threshold. Each pixel is converted to a binary form of presentation in the following manner: If the brightness is more than the threshold, 1 is obtained; if the brightness is less than the threshold, 0 is obtained. The binary code of the window is formed from the left upper corner, as presented in Figure 1. Sometimes, the mean value of the window brightnesses is selected as a threshold for binary code calculations. LBP describes the micropattern on the image. It is possible to build an LBP histogram that is computed over the whole image. Such a representation encodes only the occurrences of micropatterns without any indication of their locations [15].
To use LBP for face recognition, it is necessary to divide face images into a grid of subregions. These subregions are not necessarily well aligned with facial features. Moreover, the resulting facial description depends on the chosen sizes and positions of these subregions [15]. The neighboring pixels in LBP could make different contributions to the description of the face. Careful selection of neighboring pixels could help to improve the face recognition performance [1]. In [15], researchers adopted a heuristic approach to find the best pixel sampling pairs in local regions. In [1], the authors proposed a soft method of determining the optimal neighborhood sampling strategy. They calculated the pixel difference vectors (PDV) in such way that the PDVs of images of the same person are similar, and the differences among different people are enlarged. One of the interesting conclusions of this work was that it used a local Discriminant Face Descriptor (DFD), which describes the face structures locally and precisely, and achieved better face recognition performance than global DFD [1]; the many experiments that were performed demonstrated this fact. It should be noted that with respect to computational cost, every image was divided into 49 nonoverlapping regions, each of which corresponded to a 1024-dimensional feature; therefore, the feature dimension of DFD was1024 × 49 = 50,176.
Ahonen et al. [13] introduced LBP in facial recognition with a nearest neighbor (NN) classifier.
In [1], the authors used cropped face examples from different image databases (e.g., FERET, LFW).
During the development of the LBP methodology, a large number of variations were designed to improve the performance or expand the applications; for example, ILBP (Improved LBP) and ELBP (Extended LBP) [15]. The downside of ELBP is that it significantly increases the feature dimensionality. The feature vector sometimes has a dimensionality range of 3540 to 10,620 in the case of colored images.
LBP-based features have a large dimensionality; to reduce this, this method is combined with some popular learning techniques which were developed and used for texture recognition, and then for face recognition tasks. Recent versions of UUCoLBP, RUCoLBP, and PRICoLBP were developed on the basis of LBP [4]. For example, PRICoLBP preserves pairwise rotation invariance.
The disadvantages of LBP are rarely discussed. However, LBP has sensitivity to random and quantization noise [15]. The development of a large number of LBP variations demonstrates the wish to avoid this problem and improve the performance in different applications [16,17]. However, these improvements typically increase the computational complexity.
The literature has proposed a combination of local face descriptors LBP/LDiP/LDNP with Discrete Fourier Transform (DFT) as a global face descriptor [18]. LDiP is Local Directional Pattern and LDNP is Local Directional Number Pattern. The results were obtained using the ORL database.
WLD [10] was inspired by Weber’s Law and is used as a robust local descriptor. The WLD method was tested on texture databases and demonstrated effective results. Forhuman face detection, this method also showed promising results with the use of an SVM classifier. Sometimes, the investigators used a WLD histogram for a given image [10].
Learning DFD was proposed in [1]. Traditionally, the form of such local descriptors is predefined in a hand-crafted way. This method proposes to learn a DFD in a data-driven way. A DFD introduces discriminant learning into the feature extraction process. The DFD was tested on different face databases and demonstrated improvements in the recognition results.
In this paper, we describe all of these local descriptors in detail to demonstrate the interest of scientists and engineers in the image recognition area, and in order to have examples with which to compare the advantages and disadvantages of these methods with the methods that we are proposing.
We have developed a special feature extractor and neural classifiers that we applied to different types of images, such as handwritten digit recognition, micro-object shape recognition, face recognition and other domains [19,20,21,22,23]. Different types of neural classifiers have been developed, for example, Random Threshold Classifier (RTC), Random Subspace Neural Classifier (RSC), Limited Receptive Area Classifier (LIRA classifier), etc.
The proposed feature extractor is based on the concept of random local descriptors (RLDs). RLDs are followed by an encoder that is based on the permutation coding technique, which accounts for not only the detected features, but also the position of each feature in the image, and makes the recognition process robust to small displacements. The combination of RLDs and permutation coding permits us to obtain a sufficiently general description of the image to be recognized. The code generated by the encoder is used as input data for the PCNC neural classifier.
In this article, we describe in detail the RLD and compare it with other local descriptors. We demonstrate the possibility of an RLD application for the face recognition task. From among several tasks, including face detection, face recognition, facial expression analysis, demographic classification (classification age, gender and ethnicity, based on face images) and other applications, we selected face recognition as an application for RLDs. We apply this RLD to face recognition using the different face image databases, for example, the ORL, FEI, and FRAV3D image databases.
The advantages of RLDs are that they can be easily extracted from the raw images to allow for fast processing, and they can be combined with a neural classifier to avoid computationally expensive algorithms.
The sizes and positions of RLDs can vary; they can overlap on the face image.
This paper focuses on the face recognition task. In previous studies (regarding the ORL database of faces), we proposed the inclusion of displaced face images as part of the training set and obtained a good recognition rate [22]. FEI [24] and the 2D images from the FRAV3D [25,26,27] databases fulfilled our investigation needs for the PCNC, and allowed us to obtain superior recognition rates. As we have the results of other authors for the ORL database and LBPs, we decided to repeat our experiments for ORL and RLDs in order to compare them.
An explanation of the PCNC classifier is given in Section 2. Section 3 provides a description of the ORL, FRAV3D and FEI databases, as well as the distortions that are included in the recognition process. In Section 4, we show the results of the experiments with the PCNC neural classifier in comparison with other face recognition algorithms, and present the new results of the experiments with the ORL and FEI databases. Concluding remarks are given in Section 5.

2. Permutation Coding Neural Classifier

The PCNC is meant to be a multipurpose recognition tool. It has been tested on handwritten digits, micromechanical pieces and face recognition [20,21]. Figure 2 shows the processes that take place in the PCNC classifier.
As observed in Figure 2, there are three stages in the PCNC method: preprocessing, processing and recognition. The first stage, image preprocessing, converts color images to grayscale images. Sometimes, scientists use the color images in their investigations [28]. To reduce the complexity and determine the true invariance for face recognition, images can be converted from the RGB to gray scale, as described in [29]. In this paper, we used the equation
f ( x , y ) = ( R + G + B ) / 3 ,
The gray-scale image was then processed with a median filter to obtain edges as points of interest (POIs in Figure 2).
The recognition stage includes two substages, i.e., training and recognition, for the recognition task with the PCNC classifier.
An example of extracting the POIs is demonstrated in Figure 3. The last image is used as the input image for our classifier.
The resulting image from this stage is shown in Figure 4. The original image was taken from the FEI image database [24].
In face recognition, the feature representation of a face is the key to good performance. A good representation must minimize intraperson dissimilarities and maximize the differences among different people, as well as being fast and compact.

2.1. Extractor of Features

We propose a classifier using the concept of RLD and Frank Rosenblat’s perceptron [30]. RLD works as a general feature extractor by connecting a neuron in the associative layer to a random point in the retina (input image) and calculating a brightness function of the selected point. The scheme of the neural network recognition system is shown in Figure 5.
As shown in Figure 5, the system is based on a multilayer neural network. The first S layer (sensor layer) is the input image; the second D layer contains RLD neurons (Figure 6). The A layer is an associative layer of neurons (Figure 5). The R layer is an output layer. Each of these output neurons corresponds to a recognition image class. Here, we describe the RLD structure in detail.
The RLD scheme is presented in Figure 7. RLD is constructed around points of interest (POIs). The POI is in the center of the RLD (in Figure 7, the POI is not shown).
In this study, we assume that the POIs correspond to image locals in which the surface of the pixel brightness is not plain.
We collocate the RLD around the extracted POI. In the center of the scanning windows (h × w), there are two auxiliary rectangles: an internal rectangle with the area I × I = I2 pixels, and an external rectangle with the area E × E = E2 pixels (Figure 7). All of the pixels of the internal rectangle are connected with the neuron, with connections that have positive values for the weight being represented by wI. All of the pixels of the external rectangle are connected to the neuron with the negative-valued weights, wE. The weights wI and wE are selected according to the equation:
I 2 w I = E 2 | w E | ,
The neuron calculates the input excitation:
E i n = i = 1 I j = 1 I b i j w I i = 1 E j = 1 E b i j | w E | ,
where b i j is the brightness of the pixel that has coordinates (i,j).
The neuron output equals 1 if
| E i n | T ,
where T is the threshold. Otherwise, the neuron output equals 0.
For every RLD (Figure 7), two types of neurons are considered, similar to natural neural networks, namely, ON and OFF neurons (ON neurons correspond to pixels with a connection with the arrow or positive point; and OFF neurons correspond to pixels with a connection with the circle or negative point). ON neurons respond if the input is more than the threshold, whereas OFF neurons respond if the input is less than threshold. We use binary outputs, i.e., “1” (or active) and “0” (or inactive). In an image, these neurons correspond to the positive and negative points. Figure 8 presents an example of the RLD that determine each feature, which exists only when all of the ON and OFF neurons are active.
The ON neuron has an output of “1” if the brightness bi of the corresponding pixel is higher than the neuron threshold Ti: bi ≥ Ti.
The OFF neuron has an output of “1” if the brightness bi of the corresponding pixel is less than the neuron threshold Ti: bi < Ti.
The threshold values are randomly selected from among the brightness values Tmin ≤ Ti ≤ Tmax of the input image.
A D layer neuron (Figure 7) is a neuron that simulates a conjunction operation. It has an output of “1” if and only if all eight neurons (for connections with the arrow and four connections with circles) have outputs of “1”.
Each neuron of the dij plane (Figure 6) corresponds to the pixel that is located at the center of the I-rectangle (Figure 7).
All of the neurons that have an output of“1” are considered to be active neurons. We consider that the feature exists only if all of the positive and negative points are active; otherwise, it is absent.
All of the neurons of the associative A layer have trainable connections with R layer neurons (Figure 9). The training process is realized between these two layers by changing the weight of every connection between the A and R layers. If the answer is correct, nothing has been done. In the case of an incorrect answer, all connection weights to the incorrect neuron are reduced, and all weights to the correct neuron are increased.

2.2. Feature Encoder

To explain the feature encoder, we must introduce the following variables. Let a feature Fi be
F i = ( U i , ( P 1 i , P 2 i , , P k i ) ) ,
where Pj is the position of the feature Fi in the image. We can have the same feature in different places of the image (for example, in Figure 10a, two white lines with the same inclination, or in Figure 10b, two different pairs of feature frames from the man’s lips).
Thus, for every feature, we can define the following:
P j = ( x j ,   y j ,   1 C ( O N c , T h O N c ) ,   1 L O F F l , T h O F F l ) ,
where xj and yj are coordinates of the window of size (w × h) with a center in the point of interest, 1 C ( O N c , T h O N c ) is the conjunction function of C ON-neurons ( O N c is the position of an ON-neuron randomly generated in window (w × h)), T h O N c is the threshold of the c-th ON-neuron, 1 L O F F l , T h O F F l is the conjunction function of L OFF-neurons ( O F F l is the position of an OFF-neuron randomly generated in window (w × h)), and T h O F F l is the threshold of the l-th OFF-neuron. A feature exists in the position (xj, yj) if the result of the conjunction for the ON and OFF neurons is 1. If one of these functions is 0, then the feature does not exist in that position.
To address all of these variables as O N c and O F F l neuron positions, the thresholds T h O N c and T h O F F l for the ON and OFF neurons are randomly selected.
The coordinates xj and yj of the center point of the window (w × h) are defined in the following manner. We scan the image with this window with a scan step of one pixel. If the central point is not a point of interest, we continue to scan the image. If the central point of the window is a point of interest, then we define the feature using Equation (7). If the result is 1 (i.e., all ON and OFF neurons have given an answer), then we know that the feature exists in this position. If the result is 0 (i.e., if at least one of the ON and OFF neurons does not give the answer), then we know that the feature is absent and we must continue to scan the image with the window.
For each extracted feature Fi, the encoder creates an auxiliary binary vector or mask, which is represented as follows:
U i = ( u i 1 ,   u i 2 ,   ,   u i N ) ,
where uii is equal to 0 or 1. Ui is the feature mask vector of dimension N with K 1’s, whose initial position is randomly chosen, where K<<N (we worked with K = 16 and N = 64,000). The mask corresponds to the feature in the initial position in the image and is constant throughout the lifetime of the PCNC. The other positions of the feature are encoded with permutations of the mask. Furthermore, we will describe in detail the permutation procedure. Next, we want to terminate the description of the coding procedure. As a result of the permutation process, a new vector U i ( P z i ) is created. To code the presence of a feature in the image, we apply the disjunction operation to join all of the binary vectors of this feature in different places.
If Z be in different positions for the same feature Fi, the binary code will be
U i = z = 1 Z U z ( P z i ) ,
where the vector U z is the feature mask binary vector of size N and P z i is the position of feature i that defines the permutations of the feature mask vector; thus, U z ( P z i ) is the result of the feature mask vector permutation. N is a very large value.
Next, we explain the process of the permutation. The problem is to generate binary codes with special characteristics, i.e., the correlation between two binary vectors is a function of the distance between these two vectors. Thus, the permutation not only permits us to generate a unique code for every feature in its position, but also gives us the opportunity to analyze its correlation.
The number of permutations depends on the feature location in the image. Once the permutations of the binary vector Uz are completed, a new vector U z ( P z i ) is created.
To code the position of the feature characteristic Fi, we must define the correlation distance Dc, which is the measurement between the feature distances (in our work, we use 16 pixels).
We are given two position points, P1(x1, y1) and P2(x2, y2), for feature Fi, the vectors U 1 ( P 1 i ) and U 2 ( P 2 i ) , which code the feature for every position point, and dx, the Euclidian distance between P1 and P2 in X and dy in Y.
d x = a b s ( x 1 x 2 ) ;   d y = a b s ( y 1 y 2 ) ,
The vectors U 1 ( P 1 i ) and U 2 ( P 2 i ) are correlated if dx < Dc or dy < Dc; otherwise, there is no correlation.
To code the feature Fm position, distance Dc is predefined, and the following values must be calculated:
X = x i j / D c ,   Y = y i j / D c .
We calculate the integer parts
E ( X ) = ( int )   X ,   E ( Y ) = ( int )   Y .
The integer parts correspond to the number of complete permutations of the feature mask binary vector. To evaluate the number of partial permutations, we must calculate the fractional parts of the feature coordinates. We obtain fractional parts from the following equations:
R ( X ) = x i j E ( X ) D c ,   R ( Y ) = y i j E ( Y ) D c ,
P x = int ( R ( X ) N / D c ) ,   P y = int ( R ( Y ) N / D c ) ,
where E(X) and E(Y) are the integer parts of X and Y; R(X) and R(Y)are the fractional parts of X and Y; yij is the vertical coordinate of the detected feature; xij is the horizontal coordinate of the detected feature; N is the number of neurons; E(X) and E(Y) show the number of permutations to perform in the X and Y directions; and Px and Py are the number of neurons in the range [0, N) for which an additional permutation is needed. The value of N is changed according to the problem complexity. In the case of facial recognition, we used 65,000 neurons.
The example permutation scheme for the X coordinate is presented in Figure 11 (we select E(X) = 2, and Px = 2). The U vector is the binary vector. To more easily explain the permutation process, we use the letters that are contained in each element of the U vector. The process is performed as follows: each element from the first line is connected to a free and randomly selected element from the third line, using the permutation scheme presented in the second line (the A line). For example, 0 < −2 (0 “left arrow” 2) means that letter c from U(2) must be transported to U(0), and the scheme of the A line is used for all of the elements of the U vector. The third line is the result of the first permutation. The B line defines the second permutation, and the result is presented in the fourth line. The process repeats until all of the elements from the A line and B line end up with a one-to-one connection. In line C, only the two first elements have permutations (U(0) and U(1)), and the remaining elements do not change.
The full process of the feature extraction is shown in Figure 12.
The result of the feature extraction process is the associative vector A (binary vector or code), which equals a bitwise disjunction of all of the permutated vectors.
A = i U i .

3. Database and Distortions

3.1. ORL Database

One of the first databases for face recognition was ORL (Olivetti Research Laboratory), now administered by AT&T Laboratories Cambridge [31] (Figure 13). It has ten different images of each of 40 distinct subjects, and the original size of each image is 92 × 112 pixels. For some subjects, the images were taken at different times, varying the lighting, facial expressions (open/closed eyes, smiling/not smiling) and facial details (glasses/no glasses).
In the literature, many results were obtained using this dataset. We therefore used it to compare our results with those obtained via other methods.

3.2. FEI Database

FEI is a Brazilian face database that contains a set of face images taken at the FEI Artificial Intelligence Laboratory in São Bernardo do Campo, São Paulo, Brazil [24]. There are 14 images for each of 200 individuals, 11of which (for each individual) were taken against a white homogeneous background in an upright frontal position with a profile rotation of up to approximately 180 degrees (Figure 14). The scale might vary by approximately 10%, and the original size of each image is 640 × 480 pixels. The database mainly comprises images of students and staff at FEI, who were between 19 and 40 years of age; each has a distinct appearance, hairstyle, and adornments. The number of male and female subjects is exactly the same, i.e., 100 of each [24].
Examples (i.e., 12 images for one person) from the FEI image database are presented in Figure 15.
The FEI image database contains 2800 images (14 variants for each of 200 individuals) [24]. Several people have closed eyes (Figure 16a), and sometimes, the images have poor contrast. In Figure 16b, we present an example of a blurry image.
There are several cases in the FEI image database when the same person has one photo with glasses and another without them (Figure 17).
This FEI image database was tested with different recognition methods [32,33,34]. We selected the FEI image database with all of its imperfections to test our classifier.
The PCNC classifier with RLDs was investigated on the FRAV3D face image database [20,25,26], which included distortions, i.e., rotations.

3.3. FRAV 3D Image Database

The FRAV3D image database contains 105 subjects, mainly young adults, with approximately one woman for every three men [25,26]. There are 16 captures per person with different face expressions and/or lighting conditions.
Images from the FRAV3D database are presented in Figure 18 [25,26]. The FRAV3D face image database is one of the few 3D databases [25,26,27]. We selected 2D images for our experiments.
We trained and tested PCNC with three different databases: ORL, FRAV3D and FEI. For FRAV3D specifically, we added new rotated images to the training set to improve the recognition rate. Therefore, our present work includes the recognition of faces with PCNC while considering distortions such as displacements of images to a few pixels to the right, left, up or down, and rotations over the Y and Z axis. In other words, we tested the PCNC classifier under unconstrained situations.
Displacement distortions were taken from previous studies [20] and were added during the training session. Each new position of the initial image that was produced by distortions is considered to be an independent new image. For the experiments, we used fifteen cases: 1. initial position; 2. left shift with delta pixels; 3. left shift with 2 × delta pixels; 4. right shift with 2 × delta pixels; 5. shift up with delta pixels; 6. right shift with 3 × delta pixels; 7. left shift (4 × delta pixels); 8. shift down (delta pixels); 9. left shift (3 × delta pixels); 10. right shift (4 × delta pixels); 11. shift up (2 × delta pixels); 12. shift down (2 × delta pixels); 13. shift up (3 × delta pixels); 14. shift down (3 × delta pixels); 15. shift up (4 × delta pixels). In Section 4, we will describe the results obtained with different numbers of distortions.
We used rotation for the displacement of the images. Each pixel has a coordinate pair (x,y) that describes its position on two orthogonal axes from defined origin 0; rotation will be given around this origin. We consider the middle of the face image to be the origin O(w/2,h/2). For our experiments, we selected three values of the clockwise rotation angle (the reference point is the vertical axis) θ = 5°, 10°, 15°, and three values for counter-clockwise rotations θ = −5°, −10°, −15° [35,36].
In the case of the rotations, the RLD structure was the same, but the pixels had displacements. Near the center of the rotation, the changes were smaller than in the peripheral points. Thus, the RLDs for rotation could be useful in the training process. We present our analysis of their influence in the following chapter.

4. Experiments and Results

To investigate the RLD and compare the results with LBP or WLD, we used the ORL database.
To calculate errors, we used the equation Nerr = (M/N) × 100, where M is error responses of the PCNC and N is a total number of images. To program the PCNC and RLDs, we used the soft Visual Studio C++ (2019).
In Table 1, we present the obtained results. The first three lines were taken from [14]. That research describes nine methods and presents the recognition rates in the ORL database. We selected results which were interesting for us, i.e., from the middle of the table and the best result. Our results in all tables in this paper are presented in bold print.
The experiments with RLD demonstrate results which were comparable or better than the best results obtained in [21], and much better than those obtained using LBP and WLD. In [14], to work with LBP, the authors used the DIWT/LBP method. A detailed description of those methods is beyond the scope of this paper; rather, we simply compare our results with theirs.
It is significant to note that for every number of samples (2, 3, 4, 5), we did ten experiments and then calculated the average recognition rates, as presented in Table 2.
In Figure 19, we demonstrate the stage of recognition for 40 persons.
The results of the experiments using PCNC and RLDs were obtained with the FRAV database. As the results have already published, we only will mention them briefly here. Tests with the rotations and skewing were also performed.
The results showed that the PCNC neural classifier and the SVM [26] method suffered from the same recognition problems, i.e., rotations. On the other hand, ICP [26,37] had a lower percentage of errors due to rotations, but a larger percentage in almost all of the other tests.
Our approach is based on the addition of rotation distortions to the training set. The results improved from 46.6% to 23.00% for four distortions, from 41.7% to 21.00% for eight distortions and from 31.1% to 16.00% for 12 distortions [35,36]. In comparison with the basic version (without rotations), the new version significantly improved the recognition rate by decreasing by approximately twice the number of errors.
The FEI image database was used to test the PCNC and RLD. In Table 3, we present the results from [14] (we selected only three out of the nine methods) alongside our results obtained using the FEI image database. We selected two methods from the middle of the table (LBP and WLD) and the best result (DIWT/LBP); our result is presented in the last line. It is important to mention that we worked with the half of the FEI database. In our experiments, we used only images of 100 persons from the total of 200 in order to accelerate the investigation process. Additionally, the average recognition rate was calculated on the basis of 10 experiments for each number of samples (i.e., from three to seven). Our results are presented in bold print.
We organized two series of experiments. All images for every person (in the FEI, i.e., 14 images for each person) were divided into two groups; the first group contained images with odd numbers (Group 1), and the second group those with even numbers (Group 2). Either group could be used as a training or recognition set for the PCNC.
The first experiment used Group 2 for the PCNC training and Group 1 for the PCNC test. The second experiment used Group 1 for the PCNC training and Group 2 for the PCMC test. In both experiments, we used the distortions in the original images. For the FEI image database, we used 15 image distortions (Table 4).
If the distortion number was 1, then we used the original image (a more detailed description of distortions is presented in Paragraph 3.3). If the distortion number was 13, then we used the original image (position 1) for training, as well as the image shifted upward for 3 × ∆ pixel. In our experiments, = 4 for 15 distortions.
In Table 5, the results of two experiments are presented for different distortion numbers. Experiment 1 included Group 2 images for training and Group 1 images for testing. The best result of these experiments was5.83% error for nine distortions. Experiment 2 included Group 1 images for training and Group 2 images for testing. The best result was 14.1% for 15 distortions.
The first experiment showed better results in comparison with the second. In the second experiment, we obtained the worst result due to the last image. The brightness of the image with number 14 was very low, causing a poor image recognition result.
All of these experiments were made for a window size of (13 × 13) pixels. For every RLD, we used 3 positive and 3 negative points, which was a basic variant of the RLD structure. An example of the experiment is shown in Table 6.
Next, we investigated the influence of the number of positive and negative points on the RLD formation and the influence of the window size on the recognition rate.
Each result was evaluated as an average of five experiments to decrease the influence of the randomly selected parameter values.
Table 7 shows that the mean number of errors depends on the number of positive and negative points in the RLD. The best results were obtained when the numbers of positive and negative points each equaled two. The worst results were obtained in the cases of 4 positive points and 4 negative points and 1 positive and 1 negative point. The error rate was almost independent of the RLD window size in the range of 7 × 7 to 13 × 13 pixels.
With the same RLDs, we investigated the error number depending the training cycle number. Figure 20 demonstrates the improvements in recognition with an increase in the cycle number. With 100 training cycles, the recognition rate was 98.5%.
In Table 8, we present the recognition statistics for different numbers of classes but with the same number of RLDs. In this case, we used 400 features (RLDs).
The time of coding and recognition processes for PCNC with RLDs are presented in Figure 21.
The experiments with the PCNC, RLDs and the FEI image dataset yieldedgood results.
In this article, we investigated RLDs and compared their influence on image recognition with LBP algorithms.

5. Conclusions

The PCNC neural classifier is based on the RLD descriptors that are used for feature extraction from the image. We compared our method with the LBP principle often used in facial recognition. The PCNC neural classifier presented good results in facial recognition using the ORL, FRAV3D and FEI image databases. We demonstrated that our system yielded results better than those obtained using other methods, especially LBP. Our RLDs have advantages, because they require less time for calculations, and the number of RLDs is less than that required for LBPs. To obtain these results, we generated additional images for the training set using image distortions. First, the simplest distortions, including image displacements, were investigated. . Sometimes, we obtained twice the decrease in error for the FRAV3D image database. It was done many experiments with RLDs. It was shown that the RLD parameters have an influence on face recognition quality. We showed that RLD is a good alternative to LBPs, and that our model is a good approach for face feature detection using local descriptors. The PCNC with the RLDs demonstrated the best recognition rate, i.e., 97.49%, in comparison with 90.49% for LBPs for the ORL image database. For the FEI image database, we obtained the best recognition rate, i.e., 93.57%, in comparison with 66.74% for the LBPs. For FRAV3D, we improved the recognition rate by decreasing by approximately twice the number of errors using the RLDs and rotation of original images..

Author Contributions

Conceptualization and methodology, E.K. and T.B.; software, T.B. and A.C.; validation, A.C.; investigation, E.K.; writing—original draft preparation, T.B. and A.C.; supervision, E.K.; funding acquisition, E.K. All authors have read and agreed to the published version of the manuscript and have contributed substantially to the work reported.

Funding

This research was funded by UNAM-DGAPA-PAPIIT-IT102320.

Acknowledgments

The authors wish to thank the scientists of the University Rey Juan Carlos, Madrid, Spain, for the FRAV3D face image database, the scientists of the Department of Electrical Engineering, FEI, São Paulo, Brazil, for the FEI face image database and AT&T Laboratories, Cambridge for the ORL database.The authors express their thanks the students of the UNAM Cruz Monterrosas, Z.; Aparicio Hernandez, A.; Martínez Valdés, D.F.; Gari Santesteban, S.L.; Mosco Luciano, J.J;. Perez Robles, N.A.; Vázquez Silva, E.J.; for their contributions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Z.; Pietikainen, M.; Li, S.Z. Learning Discriminant Face Descriptor. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 289–302. [Google Scholar] [CrossRef] [PubMed]
  2. Er, M.J.; Wu, S.; Lu, J.; Toh, H.L. Face Recognition with Radial Basis Function (RBF) Neural Networks. IEEE Trans. Neural Netw. 2002, 13, 697–710. [Google Scholar]
  3. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face Recognition: A Convolutional Neural-Network Approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Qi, X.; Xiao, R.; Li, C.G.; Qiao, Y.; Guo, J.; Tang, X. Pairwise Rotation Invariant Co-occurrence Local Binary Pattern. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2199–2213. [Google Scholar] [CrossRef]
  5. Liu, W.; Wen, Y.; Yu, Z. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  6. Deng, J.; Guo, J.; Xue, N. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  7. Wang, H.; Wang, Y.; Zhou, Z. Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5265–5274. [Google Scholar]
  8. Shepley, A.J. Deep Learning for Face Recognition: A Critical Analysis, Cornell University. arXiv 2019, arXiv:1907.12739. [Google Scholar]
  9. Alghaili, M.; Li, Z.; Ali, H.A.R. FaceFilter: Face Identification with Deep Learning and Filter Algorithm, Intelligent Decision Support Systems Based on Machine Learning and Multicriteria Decision-Making. Sci. Program. 2020, 2020, 7846264. [Google Scholar] [CrossRef]
  10. Chen, J.; Shan, S.; He, C.; Zhao, G.; Pietikainen, M.; Chen, X.; Gao, W. WLD: A Robust Local Image Descriptor. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1705–1720. [Google Scholar] [CrossRef]
  11. Liu, L.; Fieguth, P.W. Texture Classification from Random Features. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 574–586. [Google Scholar] [CrossRef] [PubMed]
  12. Lowe, D. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  13. Ahonen, T.; Hadid, A.; Pietikäinen, M. Face Recognition with Local Binary Patterns. In Proceedings of the ECCV 2004, 8th European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; pp. 469–481. [Google Scholar]
  14. Muqeet, M.A.; Holambe, R.S. Local binary patterns based on directional wavelet transform for expression and pose-invariant face recognition. Appl. Comput. Inform. 2019, 15, 163–171. [Google Scholar] [CrossRef]
  15. Huang, D.; Shan, C.; Ardabilian, M.; Wang, Y.; Chen, L. Local Binary Patterns and its Application to Facial Image Analysis: A Survey. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2011, 41, 765–781. [Google Scholar] [CrossRef] [Green Version]
  16. Wolf, L.; Hassner, T.; Taigman, Y. Descriptor Based Methods in the Wild. In Proceedings of the Faces in Real-Life Images workshop at the European Conference on Computer Vision (ECCV), Marseille, France, 17–18 October 2008; pp. 1–14. [Google Scholar]
  17. Maturana, D.; Mery, D.; Soto, A. Learning Discriminative Local Binary Patterns for Face Recognition. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition (FG), Santa Barbara, CA, USA, 21–25 March 2011; pp. 470–475. [Google Scholar]
  18. Saragih, R.A.; Sudiana, D.; Gunawan, D. Combination of DFT as Global Face Descriptor and LBP/LDiP/LDNP as Local Face Descriptor for Face Recognition. J. Telecommun. Electron. Comput. Eng. 2018, 10, 99–102. [Google Scholar]
  19. Baidyk, T.; Kussul, E.; Makeyev, O.; Caballero, A.; Ruiz, L.; Carrera, G.; Velasco, G. Flat image recognition in the process of microdevice assembly. Pattern Recognit. Lett. 2004, 25, 107–118. [Google Scholar] [CrossRef]
  20. Kussul, E.; Baidyk, T. Improved Method of Handwritten Digit Recognition Tested on MNIST Database. Image Vis. Comput. 2004, 22, 971–981. [Google Scholar] [CrossRef]
  21. Baidyk, T.; Kussul, E.; Makeyev, O. Texture Recognition with Random Subspace Neural Classifier. WSEAS Trans. Circuits Syst. 2005, 4, 319–325. [Google Scholar]
  22. Kussul, E.; Baidyk, T.; Wunsch, D.; Makeyev, O.; Martín, A. Permutation Coding Technique for Image Recognition Systems. IEEE Trans. Neural Netw. 2006, 17, 1566–1579. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Kussul, E.; Baidyk, T.; Makeyev, O.; Martín, A. Image Recognition Systems Based on Random Local Descriptors. In Proceedings of the IEEE International Joint Conference on Neural Network, Vancouver, BC, Canada, 16–21 July 2006; pp. 4722–4727. [Google Scholar]
  24. FEI Face Database, Image Processing Laboratory, Department of Electrical Engineering, Centro Universitario da FEI, São Bernardo do Campo, São Paulo, Brazil. Available online: http://fei.edu.br/~cet/facedatabase.html (accessed on 8 May 2021).
  25. FRAV3D, Universidad Rey Juan Carlos. Available online: http://www.frav.es/ (accessed on 8 May 2021).
  26. Conde, C. Verification Facial Multimodal: 2D y 3D. Ph.D. Thesis, Face Recognition and Artificial Vision Group (FRAV), URJC, Madrid, Spain, 2006. [Google Scholar]
  27. Walkden, J.; Pears, N. The Utility of 3D Landmarks for Arbitrary Pose Face Recognition. 2010. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.696.8408&rep=rep1&type=pdf (accessed on 6 June 2021).
  28. Sun, Y.; Wang, X.; Tang, X. Deep Learning Face Representation from Predicting 10,000 Classes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1891–1898. [Google Scholar]
  29. Wang, W.; Wang, W.F. A Gray-Scale Face Recognition Approach. In Proceedings of the 2008 Second International Symposium on Intelligent Information Technology Application, Shanghai, China, 20–22 December 2008; pp. 395–398. [Google Scholar]
  30. Rosenblatt, F. Principles of Neurodynamics; Spartan: Washington, DC, USA, 1962. [Google Scholar]
  31. ORL Database. Available online: http://cam-orl.co.uk/facedatabase.html (accessed on 8 May 2021).
  32. Cherifi, D.; Radji, N.; Nait Ali, A. Effect of Noise, Blur and Motion on Global Appearance Face Recognition Based Methods Performance. Int. J. Comput. Appl. 2011, 16, 4–13. [Google Scholar] [CrossRef]
  33. Hashim, A.N.; Hussain, Z.M. Local and Semi-Global Feature-Correlative Techniques for Face Recognition. Int. J. Adv. Comput. Sci. Appl. 2014, 5, 157–167. [Google Scholar]
  34. El-Sayed, R.S.; El Nahas, M.Y.; El Kholy, A. Sparse Representation Approach for Variation Robust Face Recognition Using Discrete Wavelet Transform. Int. J. Comput. Sci. Issues 2012, 9, 275–280. [Google Scholar]
  35. Baidyk, T.; Kussul, E.; Cruz Monterrosas, Z.; Ibarra Gallardo, A.J.; Roldán Serrato, K.L.; Conde, C.; Serrano, A.; Martín de Diego, I.; Cabello, E. Face Recognition using a Permutation Coding Neural Classifier. Neural Comput. Appl. 2016, 27, 973–987. [Google Scholar] [CrossRef]
  36. Cruz Monterrosas, Z.; Baidyk, T.; Kussul, E.; Ibarra Gallardo, A.J. Rotation Distortions for Improvement in Face Recognition with PCNC. In Proceedings of the IEEE 3rd International Conference and Workshop on Bioinspired Intelligence, Liberia, Costa Rica, 16–18 July 2014; pp. 50–55. [Google Scholar]
  37. Cook, J.; Chandran, V.; Sridharan, S.; Fookes, C. Face Recognition from 3D Data using Iterative Closest Point Algorithm and Gaussian Mixture Models. In Proceedings of the 2nd International Symposium on 3D Data Processing, Visualization and Transmission, Thessaloniki, Greece, 9 September 2004; pp. 1–8. [Google Scholar]
Figure 1. Example of abasic LBP operator.
Figure 1. Example of abasic LBP operator.
Electronics 10 01358 g001
Figure 2. PCNC structure.
Figure 2. PCNC structure.
Electronics 10 01358 g002
Figure 3. Example of the application of a linear filter.
Figure 3. Example of the application of a linear filter.
Electronics 10 01358 g003
Figure 4. Example image that shows POIs.
Figure 4. Example image that shows POIs.
Electronics 10 01358 g004
Figure 5. Structure of the general-purpose image recognition system.
Figure 5. Structure of the general-purpose image recognition system.
Electronics 10 01358 g005
Figure 6. D layer structure.
Figure 6. D layer structure.
Electronics 10 01358 g006
Figure 7. RLD1 structure.
Figure 7. RLD1 structure.
Electronics 10 01358 g007
Figure 8. Points of interest selected by the feature extractor.
Figure 8. Points of interest selected by the feature extractor.
Electronics 10 01358 g008
Figure 9. Connections between A and R layers.
Figure 9. Connections between A and R layers.
Electronics 10 01358 g009
Figure 10. Same features in different places: (a) White lines with the same inclination; (b) Differentpairs of feature frames from the man’s lips.
Figure 10. Same features in different places: (a) White lines with the same inclination; (b) Differentpairs of feature frames from the man’s lips.
Electronics 10 01358 g010
Figure 11. Example permutation pattern for the U vector.
Figure 11. Example permutation pattern for the U vector.
Electronics 10 01358 g011
Figure 12. Feature extraction process.
Figure 12. Feature extraction process.
Electronics 10 01358 g012
Figure 13. Example from the ORL face database.
Figure 13. Example from the ORL face database.
Electronics 10 01358 g013
Figure 14. Camera positions.
Figure 14. Camera positions.
Electronics 10 01358 g014
Figure 15. Example of images (FEI face database).
Figure 15. Example of images (FEI face database).
Electronics 10 01358 g015
Figure 16. Example with closed eyes (a) and of a blurry image (b) from the FEI database.
Figure 16. Example with closed eyes (a) and of a blurry image (b) from the FEI database.
Electronics 10 01358 g016
Figure 17. Images of people with and without glasses.
Figure 17. Images of people with and without glasses.
Electronics 10 01358 g017
Figure 18. Example from the FRAV3D image database.
Figure 18. Example from the FRAV3D image database.
Electronics 10 01358 g018
Figure 19. Example of face recognition for the ORL database.
Figure 19. Example of face recognition for the ORL database.
Electronics 10 01358 g019
Figure 20. Error number as a function of training cycle number.
Figure 20. Error number as a function of training cycle number.
Electronics 10 01358 g020
Figure 21. Coding and recognition time for PCNC.
Figure 21. Coding and recognition time for PCNC.
Electronics 10 01358 g021
Table 1. Average recognition rates for the ORL face database (%).
Table 1. Average recognition rates for the ORL face database (%).
Number of Samples/
Methods
2345
LBP72.9480.6186.3990.49
WLD75.7384.689.3792.5
DIWT/LBP83.3588.2694.1797.0
RLD87.7194.1696.197.49
Table 2. Ten experiments with RLD (ORL face database (%)).
Table 2. Ten experiments with RLD (ORL face database (%)).
Number of Training Samples/Run2345
186.7988.9897.62100
290.3691.8495.2497.71
379.2995.1096.1996.0
492.8695.9294.7694.86
593.5796.3395.7197.14
685.7193.8897.1497.14
788.5791.8496.1997.14
887.595.196.6798.86
989.2995.5197.1497.14
1083.2197.1494.2998.86
Average recognition rate87.7194.1696.197.49
Table 3. Recognition rates for different methods using the FEI database (%).
Table 3. Recognition rates for different methods using the FEI database (%).
Number of Samples/
Methods
34567
LBP43.251.6456.262.7866.74
WLD5260.8264.471.7275.49
DIWT/LBP58.565.6668.477.3382.25
RLD79.6986.7288.4592.6293.57
Table 4. Errors for the FEI image database.
Table 4. Errors for the FEI image database.
Distortion NumberExperiment 1
Errors
Experiment 2
Errors
Mean%Mean%
1372410
1481511
1593612
Table 5. Results of two experiments.
Table 5. Results of two experiments.
Distortion NumberExperiment 1
Errors
Experiment 2
Errors
Mean%Mean%
162.68.94112.816.1
344.86.4107.815.4
940.85.83100.414.3
1542.46.0698.414.1
Table 6. Error of each experiment.
Table 6. Error of each experiment.
Run12345Mean%
Error2932292827294.14
Table 7. Average resukts.
Table 7. Average resukts.
Window Size
(Pixels)
Positive
Point
Negative
Point
Mean ErrorsMean
Errors (%)
13 × 1333324.57
13 × 1344466.57
13 × 1322304.29
11 × 112229.44.2
11 × 1111466.57
9 × 922294.14
7 × 72229.24.17
5 × 52224.43.49
3 × 32219.22.74
3 × 311253.57
Table 8. Recognition.
Table 8. Recognition.
Number of ClassesRecognition Rate (%)
5100
1098.57
1599.04
2099.28
2598.85
3099.04
3598.77
4098.57
4598.41
5097.71
10096.85
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Curtidor, A.; Baydyk, T.; Kussul, E. Analysis of Random Local Descriptors in Face Recognition. Electronics 2021, 10, 1358. https://doi.org/10.3390/electronics10111358

AMA Style

Curtidor A, Baydyk T, Kussul E. Analysis of Random Local Descriptors in Face Recognition. Electronics. 2021; 10(11):1358. https://doi.org/10.3390/electronics10111358

Chicago/Turabian Style

Curtidor, Airam, Tetyana Baydyk, and Ernst Kussul. 2021. "Analysis of Random Local Descriptors in Face Recognition" Electronics 10, no. 11: 1358. https://doi.org/10.3390/electronics10111358

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop