Next Article in Journal
Non-Intrusive Fish Weight Estimation in Turbid Water Using Deep Learning and Regression Models
Previous Article in Journal
Effect of Electrode Thickness on Quality Factor of Ring Electrode QCM Sensor
 
 
Correction published on 30 October 2023, see Sensors 2023, 23(21), 8825.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Face Spoofing, Age, Gender and Facial Expression Recognition Using Advance Neural Network Architecture-Based Biometric System

1
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Vijaywada 522302, India
2
Department of IT, Neil Gogte Institute of Technology, Kachawanisingaram Village, Hyderabad 500039, India
3
College of Computing Sciences and IT, Teerthanker Mahaveer University, Moradabad 244001, India
4
Department of Media and Educational Informatics, Faculty of Informatics, Eötvös Loránd University, 1053 Budapest, Hungary
5
National Research and Development Institute for Cryogenic and Isotopic Technologies—ICSI Rm, 240050 Ramnicu Valcea, Romania
6
Doctoral School, Polytechnic University of Bucharest, 313 Splaiul Independentei, 060042 Bucharest, Romania
7
Department of Power Engineering, “Gheorghe Asachi” Technical University of Iasi, 700050 Iasi, Romania
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(14), 5160; https://doi.org/10.3390/s22145160
Submission received: 22 May 2022 / Revised: 2 July 2022 / Accepted: 3 July 2022 / Published: 9 July 2022 / Corrected: 30 October 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Nowadays, the demand for soft-biometric-based devices is increasing rapidly because of the huge use of electronics items such as mobiles, laptops and electronic gadgets in daily life. Recently, the healthcare department also emerged with soft-biometric technology, i.e., face biometrics, because the entire data, i.e., (gender, age, face expression and spoofing) of patients, doctors and other staff in hospitals is managed and forwarded through digital systems to reduce paperwork. This concept makes the relation friendlier between the patient and doctors and makes access to medical reports and treatments easier, anywhere and at any moment of life. In this paper, we proposed a new soft-biometric-based methodology for a secure biometric system because medical information plays an essential role in our life. In the proposed model, 5-layer U-Net-based architecture is used for face detection and Alex-Net-based architecture is used for classification of facial information i.e., age, gender, facial expression and face spoofing, etc. The proposed model outperforms the other state of art methodologies. The proposed methodology is evaluated and verified on six benchmark datasets i.e., NUAA Photograph Imposter Database, CASIA, Adience, The Images of Groups Dataset (IOG), The Extended Cohn-Kanade Dataset CK+ and The Japanese Female Facial Expression (JAFFE) Dataset. The proposed model achieved an accuracy of 94.17% for spoofing, 83.26% for age, 95.31% for gender and 96.9% for facial expression. Overall, the modification made in the proposed model has given better results and it will go a long way in the future to support soft-biometric based applications.

1. Introduction

Today’s modern life is fully automated. Automation is a technique which reduces manual work with the help of information technology. A task which has been done repetitively can be done through automation where human interaction would be less [1]. However, automation comes with security issues. Most industries use a biometric system for the identification of the individual [2,3]. When a person tries to access a particular place or system, then the question which comes into our minds is “Is this the right person to access the particular system or a place?” or “Is this person authorized to access the system or to perform the particular action” or “Does this person belong to a particular organization or state or country?” as shown in Figure 1 [4]. To answer all these questions, we have two methods: the identity of the individual and the ownership of that identity. Both methods could be breached or copied [5,6]. These methods failed because cybercrime is increasing rapidly. According to the reports, many telecommunication and banking sectors are losing a million dollars every year because cyber-attacks on the card and password-based authentication systems. To overcome these security concerns, biometric systems play an important role. The applications where biometric systems play an important role: border control, airport biometrics, financial biometrics, healthcare biometrics, mobile biometrics and many more. In this paper, we are focusing on the healthcare biometric system. Today, the health care industry is trying to adapt the concept of automated authentication of the individual who belongs to the hospital.
According to the findings of the American Department of Health and Human Services, around 112 million people were affected by health information breaches in February 2016 [5]. In most hospitals, employees can easily access the individual patient’s personal and healthcare-related details. Therefore, there is always a possibility of stealing sensitive information. This information can be used by criminals in different ways such as identity theft, blackmailing individuals and usage of credit/debit card details to commit fraudulent purchases also. In some cases, employees have even stolen the demographic and social security information which could be helpful for them to commit various types of crimes. A biometric system can be categorized into soft biometrics and hard biometrics. Hard biometrics include facial expression, irises and fingerprints, whereas soft biometrics include gender and age information, gait analysis, height, skin color, hair color, etc. Soft computing, based on single biometric information cannot deal with spoofing attacks, large datasets and unacceptable errors, and on the other hand, hard computing alone cannot be used for many applications [7]. Based on that, we proposed a system that includes facial features, gender information and age information and it will also deal with spoofing attacks [8,9,10]. In other words, the proposed system is a combination of soft and hard biometrics which associate a multi-biometric system to enhance the accuracy of the proposed system. The proposed system makes a big difference in the field of the health care system as well as insurance companies. With the help of this system, an authorized person can only access the patient and doctors’ record, which is a big milestone for the health industry [11,12,13].
The proposed system is created primarily for health care personnel, who can use this model to classify, store, share, protect and identify individuals who are member of the healthcare system, as well as cope with record duplication. In the majority of studies, researchers have utilized a single classification approach, such as age, gender, facial expression or face spoofing. In the proposed strategy, we created a hybrid model that employs all four modules for different dataset classifications which distinguishes it from existing methods. There are particular steps that distinguish the proposed technique, and they are as follows:
  • In this paper, we will use 5-layer U-Net-based architecture used for face detection because it gives better segmentation compared to other methods.
  • A spoofing technique will help to identify the unauthorized person, and if a person is authenticated, then only he or she can access the system, which makes our system more secure.
  • Classification will be done through Alex-Net-based architecture, and individuals can be classified according to age, gender and facial expressions. This architecture has the power of giving better accuracy even when the dataset is huge.
The rest of the paper is as follows: In Section 2, we will discuss the related work done by different researchers in the area of age, facial expressions and gender classification using different methods. Section 3 deals with the proposed methodology. Section 4 discusses the results and Section 5 consists of the conclusion and future scope.

2. Literature Work

An ample amount of research work has been already done by different researchers in the field of spoofing, age, gender and facial expression. Therefore, it is difficult to discuss all the methods as a single taxonomy; still, we will try to provide a quick and the best overview of the existing approach which is used in the above-mentioned fields. Bruno Peixoto et al. proposed a method that can detect spoofing from low- and high-quality images as well as recaptured and printed images. This method has a good classification rate under different illumination conditions [1]. Yaman AKBULUT et al. had proposed a deep learning-based face spoofing detection method. The proposed method is used for detection of the liveliness of the face from video and liveliness of the face detected using LRF-ELM and CNN methods. The experiments were performed on CASIA and NUAA datasets [4]. Quoc-Tin Phan et al. proposed the LDP-TOP method for the face spoofing detection method. In this method, spatial and temporal information is used for identifying face movements. Experiments were performed on CASIA, ldiap and MSU dataset, methods that are suitable for real-time processing [5]. Graham Desmon Simanjuntak et al. proposed a face spoofing detection method based on color distortion analysis. This method is able to extract color moments and histogram features and PCA is applied for dimensionality reduction and the Naïve bayse classifier is used for classification purposes [6]. Age and gender classification is a prominent area of research because of the increasing number of applications on social media platforms. The performance of the age and gender classification-based systems do not provide significant results as compared to face recognition-based systems. Table 1 provides a little glance at the literature work in this field.
A lot of research is going on in this field for better improvement in terms of accuracy and minimizing the error rate. Gil Levi et al. proposed a CNN-based method for age and gender classification. The CNN model, which was used for implementation, is comparatively shallow, which helps reduce the number of parameters and reduces the chances of over fitting of the data. The size of the training data was comparatively large which would be helpful for improving the accuracy of the model [28]. Eran Eidinger et al. provided a dataset of face images with the label of age and gender. In the proposed method, a support vector machine was used for the estimation of feature attributes which reduced the over-fitting problem also [29]. Afshin Dehghan et al. proposed a fully automated recognition system for face, emotion and gender. The proposed method is implemented using a deep convolutional neural network (DCNN) and a huge amount of dataset was collected which reduced the recognition effort in terms of time [30]. Zakariya Qawaqneh et al. also used DCNN for age estimation. In the proposed method, the system was trained for face recognition and later was used for the estimation of age, which improved the performance of the age estimation model. The proposed method overcame the over-fitting problem, as a model was already trained with a huge number of images and subjects, which leads to better performance of the system [31]. Zukang Liao et al. [32] used a local deep neural network for gender recognition. Instead of using 100 patches per image, only 9 patches per image were extracted, which reduced the training time significantly. The performance of the proposed method was 1% less than the existing age and gender classification methods, and they concluded that mouth and eye regions have more significant features for age classification and gender classification and, therefore, the eye region plays an important role [33]. Salah Eddine-Bekhouche et al. proposed an age and gender classification method using ML-LPQ (Multi-Level Local Phase Quantization) features, and SVM is used to predict age and gender and results are better than existing methods [34]. Ehsan Fazl-Ersi et al. proposed a method for age and gender classification which integrated the various features related to age and gender and it provided more visual information. Only the ROI was extracted from the image, which described the whole image properly. Classification andrecognition were done with the help of a support vector machine and this method improved the classification and recognition accuracy comparatively [35,36]. Modesto Castrillón-Santana et al. used periocular biometrics for gender classification and results were 8% less than the existing state of art method, but this paper suggested the importance of a particular area when the image is not so clear or difficult to capture [37]. The authors had combined the features of a particular face and shoulder area and used a two-stage ensemble classifier for classification purposes. Error rate reduced by 20% compared to other GC-based results [38]. Seyed Mehdi Lajevardi proposed a TPC framework for facial expression recognition which was based on color facial images. These color components provided the extra facial information and the features were extracted using log Gabor filter and the quotient method was applied for the feature selection process [39,40,41]. The classification was done through a multi-class linear discriminant analysis classifier. The proposed method worked and outperformed on low-resolution images and different illumination conditions also. This color information helped in the improvement of the emotion recognition performance of the system [42]. S L Happy et al. proposed a method for facial expression recognition using appearance-based features. Prominent features were extracted from the image which contains most of the facial information. In the proposed methods, a free facial landmark detection technique was used, which gave similar results in terms of performance but significantly better results in terms of execution time, and results were better for low-resolution images also [43]. Jiu-Cheng Xie et al. (2019) used CNN for age classification as age estimation is a multi-class classification problem, and also used the losses function for better optimization of the predicted probability distribution of each class [44,45,46]. The proposed method gave better results on wild face and controlled dataset but the results were not satisfactory in an uncontrolled environment. Jingchun Cheng et al. (2019) proposed a gender classification method based on deep learning techniques in which multiple large face patches were used. Many CNN were trained on each predefined patches which reflected the different image resolution and partial cropping of the image. Further, the decision of every CNN was combined using the voting method which gave the gender classification results accurately [47,48,49]. Every patch was evaluated properly and complementary patches were selected for voting purposes [50,51,52]. The accuracy of the proposed method was comparatively better than the existing techniques [53]. M. Duan et al. (2018) proposed the CNN2ELM method for age estimation which is an aggregation of CNN, ELM. The working of the proposed method is divided into three parts. First, different features were extracted and age grouping had been done through the ELM classification method and the estimation of age had been done through an ELM regressor. The features which were related to age were enhanced by the fusion method. An ELM classifier was used for estimating the accuracy of age groups. The accuracy of the age estimation model was better than state of the art methods [54]. Ke Zhang et al. [55] proposed a CNN and RoR-based age and gender classification approach. To overcome the problem of over-fitting, the RoR model was trained on the dataset. This combination of CNN and RoR had given better results compared to the state of the art methods and CNN-based methods. The accuracy of the system was tested on Adience and the IMDB-WIKI dataset. H. Chen et al. (2019) proposed a face anti-spoofing method based on FARCNN. In this method, a three-way classification approach was developed [56]. A conventional RNN-based method was extended using ROI pooling feature fusion, and crystal loss function was added to the original multi-tasking loss function. To handle the various illumination condition, R-LBP was used. These detectors outperformed and results were better on cross datasets also which shown the effectiveness of the proposed method. W. Sun et al. (2020) proposed a face anti-spoofing method based on FCN-DA-LSA (Fully Convolutional Network with Domain Adaptation and Lossless Size Adaptation). Domain adaption and lossless size adaption improved the accuracy rate of face spoofing detection methods.
Peng Zhihao et al. (2020) [57] proposed a smart home system based on the deep learning concept. In this model, simulation action is involved for lights and blinds to control the lights (ON/OFF). In this model, RNN had taken historical data for simulation and predicted 15 min intervals in a day for lights and blinds. Chia-Shing Tai et al. (2019), [58] proposed a real-time smart home system based on the reinforcement learning concept. This model improved the energy efficiency of smart homes by reducing 53% of the cost, 37.5% peak value and 7.05% PAR value from an average value. Shih-Hsiung Lee et al. (2019) [59] proposed an intelligent home system based on the deep learning concept. In this module, agents recognized the human face as well as pets at home for controlling appliances. The model was trained through VGG-Net architecture using the Image-Net dataset. This model achieved a 96.88% accuracy rate while testing the module. Nada A. Rasheed et al. (2018) [60] proposed texture and color bases classification and reconstruction models for the archaeological fragments with the help of a neural network concept. This module achieved 96.1% accuracy of classification as well as reconstruction of archaeological data. This means that the proposed model achieved a high success rate in the field of reconstruction of archaeological things. Yi-Ping Hung et al. (2018) proposed a simulating- based archaeological excavation in the virtual platform and achieved better results as compared to other state of the art modules [61]. Marco Galli et al. (2018) described the importance of archaeological contexts to reconstruct the original features of archaeological excavation [62]. Athas Agapiou et al. (2018) proposed a Bayesian Neural Network-based archaeological module [63]. This fusion-based module increased the accuracy, recall, precision, classification as well as reconstruction. This fusion-based module achieved a global correlation coefficient of up to 73%. Most of the authors have worked on the accuracy of the face spoofing detection method and most of the researchers had used the traditional method for detection of spoofing attacks and much less research work was done with the help of deep learning techniques and the need to focus on the dimensionality reduction of the feature vector as well accuracy of the secured biometric module. Gender and age classification is an important task in automated face analysis. Most existing approaches for gender classification use only raw/aligned face images after face detection as input. These methods exhibit fair classification ability under constrained conditions, in which face images are acquired under similar illumination with similar poses. The performances of these methods may deteriorate when face images show drastic variances in poses as routinely encountered in real-world data. Nowadays, deep learning-based methods are used for face recognition and results are better than existing methods. The same approach could be applied to age and gender classification, which may improve the accuracy of a classification and will also be helpful for the reduction of processing time in real-time applications.

3. Proposed Methodology

A novel hybrid method using deep learning technology is used in this paper to overcome the drawbacks in face spoofing, age and gender recognition. We used U-Net based architecture for Segmentation and Alex-Net for Classification. Until now, as per our knowledge, these architectures have been used for other applications and performed well but have not been used for face spoofing, age and gender recognition. Before using U-Net, we had done a small experiment. Initially, for a particular set of values of hyperparameters for all the architectures i.e., FCN, SegNet and U-Net, training and testing are performed. In this experiment, U-Net architecture performed best among all the three architectures. Therefore, we have selected U-Net architecture for further training and testing. We have trained this architecture from scratch and used all six Benchmark/Standard datasets. We performed our experiment while changing the learning rate on various values, i.e., 10−2, 5 × 10−2, 10−3 and 5 × 10−4. Finally, we got a better result when the learning rate is 10−3. In the same way, we had done our experiment while changing the batch size of the image i.e., 32, 64, 128 and 264. Finally, we got a better result when the batch size is 128. In Figure 2, the flowchart describes the proposed methodology, in which face detection, spoofing, age and gender with various facial expressions had to be conducted.
The following steps explain the proposed methodology:
  • Image Acquisition:
An image is taken as input which has to convert into a grayscale image. The converted image is then sent to the preprocessing step for image enhancement as well as noise removal. This input image is taken from benchmark datasets used or live through a webcam.
2.
Preprocessing:
Preprocessing is performed to remove distortions or other unwanted features while identifying the face part from an image and to extract the proper features of the image corresponding to detect the boundary of the face. In preprocessing, an unwanted feature is removed through the adaptive median filter and the image is resized. In preprocessing, the image is processed through different phases, such as resize, boundary detection and normalization.
3.
Segmentation:
After preprocessing, the advanced neural network-based architecture is used for proper segmentation so that the overall face detection rate will increase. The proposed work used 5-layer U- Net-based architecture as shown in Figure 3 and it is one of the fully convolutional neural networks which works with training models of data. Finally, the proposed architecture achieved very powerful segmentation results.
This architecture is parallel in structure and makes very efficient CPUs for image processing applications. It takes less than a second to segment the image. The normal convolution operation is performed using the following steps.
  • Define input size of n i n × n i n × c of 3D volume, where nin is the row and column width, c is channels.
  • Define a set of filters, f (kernels, or feature extractors to run on the above matrix), with size 3 or 5.
  • The output image of size n o u t × n o u t × f 3D volume. (nout is output image width).
U-Net consists of contracting networks with different layers as shown in Table 2, in which all the pooling operations are replaced by upsampling operators, such that the output of the network gives an image of increased resolution, and U-Net performs the classification on every pixel and generates the output with the same size as the input.
An image is read in the form of pixels, having mxnx3 and the convolution layer is formed by sets of filters. These filters are responsible for detecting the specific pattern in the given image. These filters are convolved with the input image resulting in feature maps with different values having the dimension shown in Equation (1).
(N + 2P − F)/S + 1
N = Image Dimension; P = Padding; F = Filter Size; S = Strides
Max Pooling was used in the proposed module and it selected the maximum value using filters from a set of pixel values as shown in Figure 4.
The maximum value is extracted at every stride and the remaining are deselected. After max-pooling, the dimension of the matrix is calculated by Equation (2).
(D − FS)/S + 1
D = Dimension of pooling layer; FS = Filter Size; S = Strides
After max-pooling, we applied the activation function. The activation function will perform non-linear transformation over the provided input. Finally, while training and testing, we used the Leaky ReLU function used to overcome the problem of ReLU, and Leaky ReLU function is described in Equation (3):
f ( x ) = { 0.01 x ; f o r x < 0 x ; f o r x 0
Finally, the classification layer performs classification by converting the fully connected output layer to probability for various classes. The classification layer used the SoftMax function to obtain positive values and is described in Equation (4):
S ( y i ) = e y i j = 1 K e y j
Each block in architecture takes an input and passes it through 2 convolution layers of 3 × 3 with a stride of 2 and a max-pooling size of 2 × 2 with a corresponding cropped feature map. Table 2 shows the complete description of U-Net based architecture. For each phase, it takes two convolution layers, and it may have different phases in its architecture. After applying this advanced U-Net based architecture, we achieved an excellent face detection rate on all benchmark datasets as well as in real-time as shown in Figure 5.
When detecting the face in real-time, it takes 0.5 s (less than per second). U-Net based architecture achieved 97% accuracy in real-time and high accuracy on benchmark datasets also. Once the face part is detected from an image, then the face part is cropped for further processing as shown in Figure 6. This means feature extraction.

4. Feature Extraction and Classification

After high detection of the face part, feature extraction and classification will play a vital role in the face information of the human being. That is the reason; again, one more advanced neural network (5-Layer Alex-Net architecture) was used for this process.The Alex-Net architecture consists of 5 Convolution layers and 3 fully connected layers. Alex is used to classifying the input of 1000 different classes and generates the output vector of 1000 numbers. Alex uses more than 50 Million parameters and more than 6 lakhs neurons in its implementation. Alex uses ReLu instead of the Tanh function, which makes it six times faster than the existing CNN. The convolution window of Alex is 11 × 11. The window size in the second layer will be reduced to 5 × 5 and followed by 3 × 3 in the consequent layers with max-pooling of 3 × 3 and stride of 2. The layer-wise analysis is discussed in Figure 7.
The fifth convolution layer is followed by an overlapping max-pooling layer. This layer outputs into two fully connected layers. The second fully connected layer outputs into the Softmax layer with 1000 class labels. Finally, Figure 8 represents the successful feature extraction from a live image.
After the successful feature extraction process, the last step is classification. This classification process is divided into four sections as shown below:
  • Face Spoofing
  • Age Classification
  • Gender Recognition
  • Facial Expression
In this step, first of all, Alex-Net architecture will identify that the human face whether the human face is real or fake. If it is fake, then the particular person cannot access the details of the patient or doctors; otherwise, the authorized person can access the entire details of the patient and doctors also. The author himself considers his image as an authorized person and the result of face spoofing is shown in Figure 9 and Figure 10. In the proposed paper, we used four modules with various classifications, i.e., face spoofing (real, fake), gender (male, female), age (0–2, 3–7, 8–13, 14–20, 21–36, 37–60, 60+) and facial expression (anger, disgust, fear, happiness, sadness, surprise).

5. Results

The proposed model was developed on Matlab version 2014a and the deep learning toolbox of Matlab was used for the research. For training and testing purpose, the NVIDIA Quadro P4000 graphics card was used as the GPU, which has 8-GB of RAM. For the evaluation process, we have used a 10-fold cross-validation technique to evaluate proposed models by partitioning the benchmark datasets into a training set to train the model, and a test set to evaluate it. During the classification of the human face, Alex-Net architecture is used, which requires values of different hyper-parameters to be set. This is required to obtain the optimum performance of the architecture. These hyper-parameters are epochs, learning rate and dropout and batch size. The optimized values of these hyper parameters for each problem are mentioned in Table 3.
For training of our system, we used six publicly available Benchmark Datasets. We created a combined dataset for each module and then trained our model on these datasets. We converted all the images into the input image size of U-Net architecture, i.e., 572 × 572. For every database, we used 10-fold cross-validation techniques. U-Net-based architecture plays a significant role in pixel-based segmentation and it is more effective with limited datasets images also, which is a common problem for data-scarce computer-vision tasks. Due to the max pool, the resolution of the output will increase. We use binary cross-entropy loss function as our training objective. This architecture worked well during training.
  • Early stopping if the validation loss does not improve for 10 continuous epochs.
  • Save the weights only if there is an improvement in validation loss.

5.1. Dataset Overview

For the sample, the author used his own images for the validation of the proposed work in real-time. For the evaluation of the proposed work, six publicly available benchmark datasets were used. These included NUAA and CASIA for Face Spoofing, Adience and IOG for Age, Adience and IOG for Gender and CK+ and JAFFE [64,65] for Facial Expression for a human face after signing the agreement for their respective organizations. We created a combined dataset for each module, i.e., first module for spoofing (in both the datasets, we have 14K images approximately), the second module for age and gender (in both the datasets, we have 31K approximately) and the last module for facial expression (in both the datasets, we have 11K approximately) and then trained our model on these datasets. The entire description of the benchmark datasets is shown in Table 4.

5.2. Evaluation Parameter

The overall proposed work gives better accuracy on all the standard benchmark datasets. Accuracy is the quality or state of being correct or precise. It refers to the closeness of the measured value to the standard value. The mathematic function of the accuracy is shown in Equation (5).
A c c u r a c y = T P + T N T P + T N + F P + F N
In Equation (5), TP stands for true positive which indicates the number of positive inputs correctly identified by the model. TN stands for true negative. It is the negative input correctly identified by the model. FP stands for false-positive which means the model identifies the negative input as positive. Finally, FN stands for false-negative and it identifies the positive input as negative.

5.3. Evaluation Results in Terms of Accuracy

First Module: In the first module, the proposed work is evaluated on benchmark datasets i.e., NUAA and CASIA for face spoofing. According to Table 5, it is very clear the proposed algorithm achieved better accuracy as compared to existing methodologies: 91.1% on NUAA and 92.71% on the CASIA database. The graphical representations of face spoofing results are shown in Figure 11 and Figure 12.
Second Module: In the second module, the proposed work is evaluated on benchmark datasets, i.e., Adience and IOG for Age Classification. According to Table 6, it is very clear the proposed algorithm achieved 83.26 ± 4.3% and 76.3% accuracy for Age and theAdience and IOG database, respectively. The graphical representations of face spoofing results are shown in Figure 13 and Figure 14.
Third Module: In this module, the proposed work is evaluated on benchmark datasets, i.e., Adience and IOG for gender recognition. According to Table 7, it is very clear the proposed algorithm achieved 93.01 ± 2.03 and 94.91% accuracy for Gender recognition and the Adience and IOG database, respectively. The graphical representations of face spoofing results are shown in Figure 15 and Figure 16.
Fourth Module: Finally, in the fourth module, the proposed work is evaluated for various facial expressions on benchmark datasets, i.e., CK+ and JAFFE [64,65]. According to Table 8 and Table 9, it is very clear that the proposed algorithm achieved better accuracy for all various expressions, i.e., anger, disgust, happiness, natural, sadness and surprise. The graphical representations of facial expression are shown in Figure 17 and Figure 18.

6. Conclusions

In this paper, we have proposed a model of biometric authentication based on face spoofing, facial expression, age, and gender detection which can be used in various places, but is susceptible to intrusion. The significant contribution of this paper is the use of U-Net and AlexNet architecture in collaboration. The U-Net architecture is used for segmentation purposes, and for the classification purpose, Alex-Net is applied. The proposed model is trained and tested on six benchmark datasets i.e., NUAA, CASIA, Adience, IOG, CK+ and JAFFE. Experiments have been performed for spoofing, age and gender classification and facial expression recognition. Comparative analysis has been done with existing methods. Performance of face spoofing is tested on NUAA and CASIA dataset. Accuracy is 91.1% on NUAA and 92.71% on the CASIA dataset, which is better than existing techniques. Accuracy of age classification is tested on Adience and IOG dataset, which is 83.26 ± 4.3% and 76.3%, respectively. Accuracy of gender classification is tested on Adience and IOG dataset, which is 93.01 ± 2.03% and 94.91% respectively. For facial expression recognition, we focused on anger, disgust, happiness, sadness and surprise parameters. The proposed method gives better results for all parameters except surprise. The results prove that our methods have better learning and classification performance compared to existing methods, which can be a boon for maintaining security at vulnerable places by security agencies.
In the future, the model can be trained and tested with real-time datasets and efforts can also be made to make a small prototype of the secured biometric system, which can be useful in various applications i.e., hospitality, military, banking, aviation, the insurance sector, etc. This can also be used in the healthcare sector such as in ICUs and other restricted areas where the respective surgeons/doctors can access the details of patients as well as their reports, especially since day-to-day crime is increasing in the electronic health care system.

Author Contributions

Conceptualization: S.K., S.R. and A.J.; Methodology, Conceptualization, Formal Analysis, Investigation, Validation, Writing—Original draft, Visualization and Project Administration: C.V., M.S.R., Z.I. and B.C.N.; Funding Acquisition, Project Administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Researchers Supporting by Gheorghe Asachi Technical University of Iasi, Romania.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The work of Chaman Verma and Zoltán Illés was supported by the European Social Fund under the project “Talent Management in Autonomous Vehicle Control Technologies” (EFOP-3.6.3-VEKOP-16-2017-00001). Subprogram 1.1. Institutional performance-Projects to finance excellence in RDI, Contract No. 19PFE/30.12.2021 and a grant of the National Center for Hydrogen and Fuel Cells (CNHPC)—Installations and Special Objectives of National Interest (IOSIN). This paper was partially supported by UEFISCDI Romania and MCI through projects AISTOR, FinSESco, CREATE, I-DELTA, DEFRAUDIFY, Hydro3D, FED4FIRE—SO-SHARED, AIPLAN—STORABLE, EREMI and by European Union’s Horizon 2020 research and innovation program under grant agreements No. 872172 (TESTBED2) and No. 777996 (SealedGRID).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bruno, P.; Michelassi, C.; Rocha, A. Face liveness detection under bad illumination conditions. In Proceedings of the 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3557–3560. [Google Scholar]
  2. Yang, J.; Lei, Z.; Liao, S.; Li, S.Z. Face liveness detection with component dependent descriptor. In Proceedings of the IEEE International Conference on Biometrics (ICB), Madrid, Spain, 4–7 June 2013; pp. 1–6. [Google Scholar]
  3. Jukka, K.; Hadid, A.; Pietikäinen, M. Context-based face anti-spoofing. In Proceedings of the Sixth IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS), Arlington, VA, USA, 29 September–2 October 2013; pp. 1–8. [Google Scholar]
  4. Yaman, A.; Şengür, A.; Budak, Ü.; Ekici, S. Deep learning-based face liveness detection in videos. In Proceedings of the IEEEInternational Artificial Intelligence and Data Processing Symposium (IDAP), Malatya, Turkey, 16–17 September 2017; pp. 1–4. [Google Scholar]
  5. Quoc-Tin, P.; Dang-Nguyen, D.; Boato, G.; de Natale, F.G.B. Face spoofing detection using LDP-TOP. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 404–408. [Google Scholar]
  6. Simanjuntak, G.D.; Ramadhani, K.N.; Arifianto, A. Face Spoofing Detection using Color Distortion Features and Principal Component Analysis. In Proceedings of the 7th IEEE International Conference on Information and Communication Technology (ICoICT), Kuala Lumpur, Malaysia, 24–26 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  7. Makinen, E.; Raisamo, R. Evaluation of Gender Classification Methods with Automatically Detected and Aligned Faces. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 541–547. [Google Scholar] [CrossRef]
  8. Jiang, Y.; Li, S.; Liu, P.; Dai, Q. Multi-feature deep learning for face gender recognition. In Proceedings of the 7thIEEE Joint International Information Technology and Artificial Intelligence Conference, Chongqing, China, 20–21 December 2014; pp. 507–511. [Google Scholar]
  9. Chen, H.; Wei, W. Pseudo-Example Based Iterative SVM Learning Approach for Gender Classification. In Proceedings of the 6th World Congress on Intelligent Control and Automation, Dalian, China, 21–23 June 2006; pp. 9528–9532. [Google Scholar]
  10. Kabasakal, B.; Sümer, E. Gender recognition using innovative pattern recognition techniques. In Proceedings of the 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018; pp. 1–4. [Google Scholar]
  11. Mayo, M.; Zhang, E. Improving face gender classification by adding deliberately misaligned faces to the training data. In Proceedings of the 23rd International Conference Image and Vision Computing New Zealand, Christchurch, New Zealand, 26–28 November 2008; pp. 1–5. [Google Scholar]
  12. Shabanian, M.; Eckstein, E.C.; Chen, H.; DeVincenzo, J.P. Classification of Neurodevelopmental Age in Normal Infants Using 3D-CNN based on Brain MRI. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; pp. 2373–2378. [Google Scholar]
  13. Aydogdu, M.F.; Celik, V.; Demirci, M.F. Comparison of Three Different CNN Architectures for Age Classification. In Proceedings of the 11th IEEE International Conference on Semantic Computing (ICSC), San Diego, CA, USA, 30 January–1 February 2017; pp. 372–377. [Google Scholar]
  14. Zheng, T.; Deng, W.; Hu, J. Deep Probabilities for Age Estimation. In Proceedings of the IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  15. Chen, S.; Zhang, C.; Dong, M.; Le, J.; Rao, M. Using Ranking-CNN for Age Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 742–751. [Google Scholar]
  16. Kumar, S.; Jain, A.; Kumar Agarwal, A.; Rani, S.; Ghimire, A. Object-Based Image Retrieval Using the U-Net-Based Neural Network. Comput. Intell. Neurosci. 2021, 21, 1–14. [Google Scholar] [CrossRef] [PubMed]
  17. Hollingsworth, K.; Bowyer, K.W.; Flynn, P.J. Pupil dilation degrades iris biometric performance. Comput. Vis. Image Underst. 2009, 113, 150–157. [Google Scholar] [CrossRef]
  18. Fairhurst, M.; Erbilek, M. Analysis of Physical Ageing Effects in Iris Biometrics. IET Comput. Vis. 2011, 5, 358–366. [Google Scholar] [CrossRef]
  19. Bekhouche, S.E.; Ouafi, A.; Benlamoudi, A.; Taleb-Ahmed, A.; Hadid, A. Facial age estimation and gender classification using multi level local phase quantization. In Proceedings of the 3rd International Conference on Control, Engineering and Information Technology (CEIT), Tlemcen, Algeria, 25–27 May 2015; pp. 1–4. [Google Scholar]
  20. Liu, X.; Li, J.; Hu, C.; Pan, J. Deep convolutional neural networks-based age and gender classification with facial images. In Proceedings of the First International Conference on Electronics Instrumentation and Information Systems (EIIS), Harbin, China, 3–5 June 2017; pp. 1–4. [Google Scholar]
  21. Dileep, M.R.; Danti, A. Multiple hierarchical decision on neural network to predict human age and gender. In Proceedings of the International Conference on Emerging Trends in Engineering, Technology and Science (ICE.TETS), Pudukkottai, India, 24–26 February 2016; pp. 1–6. [Google Scholar]
  22. Hu, K.; Liu, C.; Yu, X.; Zhang, J.; He, Y.; Zhu, H. A 2.5D Cancer Segmentation for MRI Images Based on U-Net. In Proceedings of the 5th International Conference on Information Science and Control Engineering (ICISCE), Zhengzhou, China, 20–22 July 2018; pp. 6–10. [Google Scholar]
  23. Hosseini, S.; Cho, N.I. Gf-capsnet: Using Gabor jet and capsule networks for facial age, gender, and expression recognition. In Proceedings of the 14th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–8. [Google Scholar]
  24. Gurnani, A.; Shah, K.; Gajjar, V.; Mavani, V.; Khandhediya, Y. SAF-BAGE: Salient Approach for Facial Soft-Biometric Classification-Age, Gender, Facial Expression. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 7–11 January 2019; pp. 839–847. [Google Scholar]
  25. Yu, Z.; Zhao, C.; Wang, Z.; Qin, Y.; Su, Z.; Li, X.; Zhou, F.; Zhao, G. Searching Central Difference Convolutional Networks for Face Anti-Spoofing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5294–5304. [Google Scholar]
  26. Lin, J.-D.; Lin, H.-H.; Dy, J.; Chen, J.-C.; Tanveer, M.; Razzak, I.; Hua, K.-L. Lightweight Face Anti-Spoofing Network for Telehealth Applications. IEEE J. Biomed. Health Inform. 2021, 26, 1987–1996. [Google Scholar] [CrossRef] [PubMed]
  27. Liu, W.; Wei, X.; Lei, T.; Wang, X.; Meng, H.; Nandi, A.K. Data Fusion based Two-stage Cascade Framework for Multi-Modality Face Anti-Spoofing. IEEE Trans. Cogn. Dev. Syst. 2022, 14, 672–683. [Google Scholar] [CrossRef]
  28. Levi, G.; Hassner, T. Age and gender classification using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 34–42. [Google Scholar]
  29. Eidinger, E.; Enbar, R.; Hassner, T. Age and gender estimation of unfiltered faces. IEEE Trans. Inf. Forensics Secur. 2014, 9, 2170–2179. [Google Scholar] [CrossRef]
  30. Dehghan, A.; Ortiz, E.G.; Shu, G.; Masood, S.Z. Dager: Deep age, gender and emotion recognition using convolutional neural network. arXiv 2017, arXiv:1702.04280. [Google Scholar]
  31. Zakariya, Q.; Mallouh, A.A.; Barkana, B.D. Deep Convolutional Neural Network for Age Estimation based on VGG-Face Model. arXiv 2017, arXiv:1709.01664. [Google Scholar]
  32. Liao, Z.; Petridis, S.; Pantic, M. Local Deep Neural Networks for Age and Gender Classification. arXiv 2017, arXiv:1703.08497. [Google Scholar]
  33. Hassner, T.; Harel, S.; Paz, E.; Enbar, R. Effective face frontalization in unconstrained images. In Proceedings of the IEEE conference on computer vision and pattern recognition, Boston, MA, USA, 7–12 June 2015; pp. 4295–4304. [Google Scholar]
  34. Misra, N.R.; Kumar, S.; Jain, A. A Review on E-waste: Fostering the Need for Green Electronics. In Proceedings of the IEEE International Conference on Computing, Communication, and Intelligent Systems (ICCCIS), Greater Noida, India, 19–20 February 2021; pp. 1032–1036. [Google Scholar] [CrossRef]
  35. Alnajar, F.; Shan, C.; Gevers, T.; Geusebroek, J.-M. Learning-based encoding with soft assignment for age estimation under unconstrained imaging conditions. Image Vis. Comput. 2012, 30, 946–953. [Google Scholar] [CrossRef]
  36. Fazl-Ersi, E.; Mousa-Pasandi, M.E.; Laganiere, R.; Awad, M. Age and gender recognition using informative features of various types. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 5891–5895. [Google Scholar]
  37. Sun, W.; Song, Y.; Zhao, H.; Jin, Z. A Face Spoofing Detection Method Based on Domain Adaptation and Lossless Size Adaptation. IEEE Access 2020, 8, 66553–66563. [Google Scholar] [CrossRef]
  38. Castrillón-Santana, M.; Lorenzo-Navarro, J.; Ramón-Balmaseda, E. On using periocular biometric for gender classification in the wild. Pattern Recognit. Lett. 2016, 82, 181–189. [Google Scholar] [CrossRef]
  39. Mery, D.; Bowyer, K. Recognition of Facial Attributes Using Adaptive Sparse Representations of Random Patches. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 778–792. [Google Scholar]
  40. Tapia, J.E.; Perez, C.A. Gender classification based on the fusion of different spatial scale features selected by mutual information from the histogram of LBP, intensity, and shape. IEEE Trans. Inf. Forensics Secur. 2013, 8, 488–499. [Google Scholar] [CrossRef]
  41. Mansanet, J.; Albiol, A.; Paredes, R. Local Deep Neural Networks for gender recognition. Pattern Recognit. Lett. 2016, 70, 80–86. [Google Scholar] [CrossRef]
  42. Lajevardi, S.M.; Wu, H.R. Facial Expression Recognition in PerceptualColor Space. IEEE Trans. Image Process. 2012, 21, 3721–3732. [Google Scholar] [CrossRef]
  43. Happy, S.L.; Routray, A. Automatic facial expression recognition using features of salient facial patches. IEEE Trans. Affect. Comput. 2014, 6, 1–12. [Google Scholar] [CrossRef]
  44. Kumar, S.; Singh, S.; Kumar, J. Face Spoofing Detection Using Improved SegNet Architecture with Blur Estimation Technique. Int. J. Biom. 2020, 13, 131–149. [Google Scholar]
  45. Kumar, S.; Singh, S.; Kumar, J. Gender Classification Using Machine Learning with Multi-Feature Method. In Proceedings of the IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; pp. 648–653. [Google Scholar]
  46. Kumar, S.; Singh, S.; Kumar, J. Live Detection of Face Using Machine Learning with Multi-feature Method. Wirel. Pers. Commun. 2018, 103, 2353–2375. [Google Scholar] [CrossRef]
  47. Kumar, S.; Singh, S.; Kumar, J. Automatic Live Facial Expression Detection Using Genetic Algorithm with Haar Wavelet Features and SVM. Wirel. Pers. Commun. 2018, 103, 2435–2453. [Google Scholar] [CrossRef]
  48. Kumar, S.; Singh, S.; Kumar, J. Multiple Face Detection Using Hybrid Features with SVM Classifier. In Data and Communication Networks; Springer: Singapore, 2019; pp. 253–265. [Google Scholar]
  49. Kumar, S.; Singh, S.; Kumar, J. A Study on Face Recognition Techniques with Age and Gender Classification. In Proceedings of the IEEE International Conference on Computing, Communication and Automation (ICCCA), Greater Noida, India, 5–6 May 2017; pp. 1001–1006. [Google Scholar]
  50. Kumar, S.; Singh, S.; Kumar, J. A Comparative Study on Face Spoofing Attacks. In Proceedings of the IEEE International Conference on Computing, Communication and Automation (ICCCA), Greater Noida, India, 5–6 May 2017; pp. 1104–1108. [Google Scholar]
  51. Kumar, S.; Singh, S.; Kumar, J. Automatic Face Detection Using Genetic Algorithm for Various Challenges. Int. J. Sci. Res. Mod. Educ. 2017, 2, 197–203. [Google Scholar]
  52. Xie, J.-C.; Pun, C.-M. Chronological Age Estimation Under the Guidance of Age-Related Facial Attributes. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2500–2511. [Google Scholar] [CrossRef]
  53. Cheng, J.; Li, Y.; Wang, J.; Yu, L.; Wang, S. Exploiting effective facial patches for robust gender recognition. Tsinghua Sci. Technol. 2019, 24, 333–345. [Google Scholar] [CrossRef]
  54. Duan, M.; Li, K.; Li, K. An Ensemble CNN2ELM for Age Estimation. IEEE Trans. Inf. Forensics Secur. 2018, 13, 758–772. [Google Scholar] [CrossRef]
  55. Zhang, K.; Gao, C.; Guo, L.; Sun, M.; Yuan, X.; Han, T.X.; Zhao, Z.; Li, B. Age Group and Gender Estimation in the Wild with Deep RoR Architecture. IEEE Access 2017, 5, 22492–22503. [Google Scholar] [CrossRef]
  56. Chen, H.; Chen, Y.; Tian, X.; Jiang, R. A Cascade Face Spoofing Detector Based on Face Anti-Spoofing R-CNN and Improved Retinex LBP. IEEE Access 2019, 7, 170116–170133. [Google Scholar] [CrossRef]
  57. Peng, Z.; Li, X.; Yan, F. An Adaptive Deep Learning Model for Smart Home Autonomous System. In Proceedings of the IEEE International Conference on Intelligent Transportation, Big Data and Smart City (ICITBS), Vientiane, Laos, 11–12 January 2020; pp. 707–710. [Google Scholar]
  58. Tai, C.-S.; Hong, J.-H.; Fu, L.-C. A Real-time Demand-side Management System Considering User Behavior Using Deep Q-Learning in Home Area Network. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 4050–4055. [Google Scholar]
  59. Lee, S.-H.; Yang, C.-S. An intelligent home access control system using deep neural network. In Proceedings of the IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), Taipei, Taiwan, 12–14 June 2017; pp. 281–282. [Google Scholar]
  60. Rasheed, N.A.; Nordin, M.J. Classification and reconstruction algorithms for the archaeological fragments. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 883–894. [Google Scholar] [CrossRef]
  61. Hung, Y.P.; Huang, T.M.; Li, K.T.; Rajapakse, R.J.; Chen, Y.S.; Han, P.H.; Liu, I.S.; Wang, H.C.; Yi, D.C. Simulating the Activity of Archaeological Excavation in the Immersive Virtual Reality. In Proceedings of the 3rd Digital Heritage International Congress (DigitalHERITAGE) held jointly with 24th IEEE International Conference on Virtual Systems and Multimedia (VSMM 2018), San Francisco, CA, USA, 26–30 October 2018; pp. 1–4. [Google Scholar]
  62. Galli, M.; Inglese, C.; Ismaelli, T.; Griffo, M. Rome under Rome: Survey and analysis of the east excavation area beneath the Basilica Iulia. In Proceedings of the 3rd Digital Heritage International Congress (DigitalHERITAGE) held jointly with 24th IEEE International Conference on Virtual Systems and Multimedia (VSMM 2018)), San Francisco, CA, USA, 26–30 October 2018; pp. 1–4. [Google Scholar]
  63. Agapiou, A.; Sarris, A. Beyond GIS Layering: Challenging the (Re)use and Fusion of Archaeological Prospection Data Based on Bayesian Neural Networks (BNN). Remote Sens. 2018, 10, 1762. [Google Scholar] [CrossRef]
  64. Lyons, M.J.; Kamachi, M.; Gyoba, J. Coding Facial Expressions with Gabor Wavelets (IVC Special Issue). arXiv 2020, arXiv:2009.05938. [Google Scholar] [CrossRef]
  65. Lyons, M.J. “Excavating AI” Re-excavated: Debunking a Fallacious Account of the JAFFE Dataset. arXiv 2021, arXiv:2107.13998. [Google Scholar] [CrossRef]
Figure 1. General diagram of Authentication [4].
Figure 1. General diagram of Authentication [4].
Sensors 22 05160 g001
Figure 2. Flow Chart of Proposed Methodology.
Figure 2. Flow Chart of Proposed Methodology.
Sensors 22 05160 g002
Figure 3. Five-Layer U-Net based Architecture.
Figure 3. Five-Layer U-Net based Architecture.
Sensors 22 05160 g003
Figure 4. Max-pooling Structure of U-Net/Alex-Net-based Architecture.
Figure 4. Max-pooling Structure of U-Net/Alex-Net-based Architecture.
Sensors 22 05160 g004
Figure 5. Real-time face detection with the processing time.
Figure 5. Real-time face detection with the processing time.
Sensors 22 05160 g005
Figure 6. Cropped Image.
Figure 6. Cropped Image.
Sensors 22 05160 g006
Figure 7. Five-Layer Alex-Net-based Architecture.
Figure 7. Five-Layer Alex-Net-based Architecture.
Sensors 22 05160 g007
Figure 8. Feature Extraction from cropped Image.
Figure 8. Feature Extraction from cropped Image.
Sensors 22 05160 g008
Figure 9. Results of Real-time face detection.
Figure 9. Results of Real-time face detection.
Sensors 22 05160 g009
Figure 10. Overall GUI of Proposed Work.
Figure 10. Overall GUI of Proposed Work.
Sensors 22 05160 g010
Figure 11. Accuracy Comparison for Face Spoofing on NUAA.
Figure 11. Accuracy Comparison for Face Spoofing on NUAA.
Sensors 22 05160 g011
Figure 12. Accuracy Comparison for Face Spoofing on CASIA.
Figure 12. Accuracy Comparison for Face Spoofing on CASIA.
Sensors 22 05160 g012
Figure 13. Accuracy Comparison for Age Classification on Adience.
Figure 13. Accuracy Comparison for Age Classification on Adience.
Sensors 22 05160 g013
Figure 14. Accuracy Comparison for Age Classification on IOG.
Figure 14. Accuracy Comparison for Age Classification on IOG.
Sensors 22 05160 g014
Figure 15. Accuracy Comparison for Gender Classification on Adience.
Figure 15. Accuracy Comparison for Gender Classification on Adience.
Sensors 22 05160 g015
Figure 16. Accuracy Comparison for Gender Classification on IOG.
Figure 16. Accuracy Comparison for Gender Classification on IOG.
Sensors 22 05160 g016
Figure 17. Accuracy Comparison for Facial Expression CK+.
Figure 17. Accuracy Comparison for Facial Expression CK+.
Sensors 22 05160 g017
Figure 18. Accuracy Comparison for Facial Expression JAFFE.
Figure 18. Accuracy Comparison for Facial Expression JAFFE.
Sensors 22 05160 g018
Table 1. Literature work on existing technology.
Table 1. Literature work on existing technology.
Ref.Author and YearDatasetPre-ProcessingSegment-ationMethodClassificationPerformance/Accuracy
[7]ErnoMakinen, 2008IIM, FERETYNGenderSVM83.38% accuracy with automatic alignment method.
[8]Y. Jiang, 2014Mixed FERET, CAS-PEAL datasetNNGenderJFLDNNsAccuracy is 89.63% on mixes dataset, which is good compared to CNN, LBP and CNN + LBP.
[9]Huajie Chen, 2006-NNGenderSVM and AAMClassifier performance has been improved by adding pseudo examples
[10]B. Kabasakal, 2018LFW, IMDB and WIKIYNGenderGoogle Net DNN, SVMThe performance of Google Net is 94.76%, which is better than SVM
[11]M. Mayo, 2008---NNGenderSVMLinear, SVMQuad, RF200The accuracy of SVMQuad is 92.5%.
[12]M. Shabanian, 2019317 MRI data from NDAYNAge3DCNNAchieved sensitivity by 99% and specificity by98.3%.
[13]M. F. Aydogdu, 2017MORPHYNAge6 layer CNN, ResNet18 and ResNet-34The performance of ResNet18 and ResNet-34 is better for the age classification problem.
[14]T. Zheng, 2017MORPHYNAgeDeep probabilities CNNProposed methods have better accuracy compared to state of theart methods.
[15]S. Chen, 2017MORPH album 2YNAgeSVM multi-class CNNThe proposed method has fewer errors in estimation compared to existing methods
[16]Fenker S.Own dataset of 630 imagesYNFeature extraction and age prediction __69% of accuracy has been achieved.
[17]Karen Hollingsworth et al., 2009Own database collected at UniversityYNAge prediction using IRIS __Achieved 70% accuracy
[18]LVQNet et al., 2011CASIA-IrisV1YNAge prediction using IRISCNNLVQNet required 31 iterations for better results
[19]Salah EddineBekhouche et al., 2015GroupsNNAge and gender classificationSVM with non-linear kernelAccuracy of age and gender classification is 88.8% 79.1%.
[20]X. Liu et al., 2017Adience, CAS-PEALNNAge and gender classificationGoogle NetAccuracy of age and gender classification is about 98%.
[21]M. R. Dileep et al., 20161000 greyscale facial imageNNAge and gender classificationFFANNAccuracy of age and gender classification is about 95%.
[22]K. Hu et al., 2018MRI imageNYsegmentationU-NetThe performance of the proposed method is higher compared to other methods
[23]Sepidehsadat Hosseini et al., 2019Web face Morph II, FG-NetNNAge and gender, facial expression classificationGF-CapsNetPerformance of Proposed GF- CapsNet is better than plain CNN for age, gender
and feature expression recognition.
[24]Ayesha Gurnani et al.Adience, AffectNet NNAge and gender, facial expression classificationSAF-BAGEPerformance is better comparatively.
[25]Zitong Yu et al., 2020OULU-NPU dataset, CASIA MFSD to Replay-Attack datasetsNNFace anti spoofingCDCN0.2% ACER in Protocol- 1 of OULU-NPU dataset,
HTER from CASIAMFSD to Replay-Attack datasets
[26]Jiun-Da Lin et al., 2022Mask DatasetNNFace anti spoofingArcFace Classifier (AC)Performance of the proposed method is better than existing systems.
[27]Weihua Liu et al., 2021MIP-2D and MIP-3D, CASIA-SURFNYFace anti spoofingD-Net, M-NetACER of 0.1071% outperforms all three comparative models with ACER of 0.4152%, 0.3425%, 0.1102%
Table 2. Convolution Layers output of U-Net based Architecture.
Table 2. Convolution Layers output of U-Net based Architecture.
PhasesInput
Image
CNN LayersFilterOutput ImageSampling TypeStride
1128 × 128 × 3216- 3 × 3128 × 128 × 16down2
2128 × 128 × 16232- 2 × 264 × 64 × 32down2
364 × 64 × 3222 × 232 × 32 × 32down2
332 × 32 × 32264- 3 × 332 × 32 × 64down2
432 × 32 × 6422 × 216 × 16 × 64down2
416 × 16 × 642128- 3 × 316 × 16 × 128down2
516 × 16 × 12822 × 28 × 8 × 128down2
58 × 8 × 1282256- 3 × 38 × 8 × 256up2
58 × 8 × 2562-16 × 16 × 128up2
516 × 16 × 1282-16 × 16 × 256up2
516 × 16 × 2562128- 3 × 316 × 16 × 128up2
616 × 16 × 128 -32 × 32 × 64upsampling to 32 × 32 × 642
632 × 32 × 64 -32 × 32 × 128 2
732 × 32 × 128264- 3 × 332 × 32 × 64 2
732 × 32 × 64 -64 × 64 × 32upsampling to 64 × 64 × 322
64 × 64 × 32 64 × 64 × 64 2
864 × 64 × 64232- 3 × 364 × 64 × 32upsampling to128 × 128 × 162
8128 × 128 × 16 -128 × 128 × 32 2
9128 × 128 × 32216- 3 × 3128 × 128 × 16128 × 128 × 12
Table 3. Details of Alex-Net Classifier.
Table 3. Details of Alex-Net Classifier.
Parameter NameSpoofingAge and GenderFacial Expression
Epochs354540
Learning Rate0.0010.0010.001
Droupout0.350.400.35
Batch Size128128128
Table 4. Details of Benchmark/Standard Datasets.
Table 4. Details of Benchmark/Standard Datasets.
Sr. No.Dataset NameRemarks
1NUAASubjects = 15, Real Image = 5000, Fake Image = 7500
2CASIASubjects = 50, Real Image = 500, Fake Image = 450
3AdienceTotal Image = 26,500, 8 Types of Categories
4IOGTotal Image = 5100, 7 Types of Categories
5CK+Subject = 123, Sequence = 593, Total Images = 10,600
6JAFFETotal Image: 213, 7 Facial Expression
Table 5. Accuracy Comparison for Face Spoofing on NUAA and CASIA.
Table 5. Accuracy Comparison for Face Spoofing on NUAA and CASIA.
Sr. No.ReferencesAccuracy on NUAAReferencesAccuracy on CASIA
1Bruno et al. [1]82.9%Bruno et al. [1]83%
2Zhen et al. [2]86.9%Zhen et al. [2]88.2%
3Jukka et al. [3]82.3%Duc-Tien et al. [5]91.1%
4Yaman et al. [4] 77.51%G. Desmon et al. [6]91.2%
5Proposed Methodology91.1%Proposed Methodology92.71%
Table 6. Accuracy Comparison for Age Classification on Adienceand IOG.
Table 6. Accuracy Comparison for Age Classification on Adienceand IOG.
Sr. No.ReferencesAccuracy on AdienceReferencesAccuracy on IOG
1Gil et al. [28]49.6 ± 4.7%A. Ouafi et al. [34]55.9%
2Eran et al. [29]44.9 ± 2.7%C. Shan et al. [35]60.1%
3Afshin et al. [30]60.9 ± 4.1%M. Awad et al. [36]62.9%
4Zakariya et al. [31]60.12%R Enbar et al. [37]67.1%
5Proposed Methodology83.26 ± 4.3%Proposed Methodology76.3%
Table 7. Accuracy Comparison for Gender Classification on Adienceand IOG.
Table 7. Accuracy Comparison for Gender Classification on Adienceand IOG.
Sr. No.MethodAccuracy on AdienceMethodAccuracy on IOG
1Gil et al. [28]86.9 ± 1.6%J. L. Navarro et al. [38]91.9%
2Eran et al. [29]78.1 ± 1.4%K. Bowyer et al. [39]92.8%
3Zukang et al. [32]79.13%C. A. Perez et al. [40]94.5%
4Tal et al. [33]79.7 ± 0.7%A. Albial et al. [41]96.7%
5Proposed Methodology93.01 ± 2.03%Proposed Methodology94.91%
Table 8. Accuracy Comparison for Facial Expression CK+.
Table 8. Accuracy Comparison for Facial Expression CK+.
ExpressionH. R. Wu et al. [42]A. Routray et al. [43]Proposed Accuracy (%)
Anger61.0988.692.7
Disgust58.6194.4196.7
Happiness74.2193.995.6
Sadness68.8195.0996.07
Surprise85.2297.1794.31
Table 9. Accuracy Comparison for Facial Expression JAFFE.
Table 9. Accuracy Comparison for Facial Expression JAFFE.
ExpressionH. R. Wu et al. [42]H. R. Wu et al. [43]Proposed Accuracy (%)
Anger63.0199.996.41
Disgust58.5187.394.13
Fear61.994.0195.43
Happiness76.4197.0196.73
Sadness68.8178.1195.07
Surprise83.9197.2696.51
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kumar, S.; Rani, S.; Jain, A.; Verma, C.; Raboaca, M.S.; Illés, Z.; Neagu, B.C. Face Spoofing, Age, Gender and Facial Expression Recognition Using Advance Neural Network Architecture-Based Biometric System. Sensors 2022, 22, 5160. https://doi.org/10.3390/s22145160

AMA Style

Kumar S, Rani S, Jain A, Verma C, Raboaca MS, Illés Z, Neagu BC. Face Spoofing, Age, Gender and Facial Expression Recognition Using Advance Neural Network Architecture-Based Biometric System. Sensors. 2022; 22(14):5160. https://doi.org/10.3390/s22145160

Chicago/Turabian Style

Kumar, Sandeep, Shilpa Rani, Arpit Jain, Chaman Verma, Maria Simona Raboaca, Zoltán Illés, and Bogdan Constantin Neagu. 2022. "Face Spoofing, Age, Gender and Facial Expression Recognition Using Advance Neural Network Architecture-Based Biometric System" Sensors 22, no. 14: 5160. https://doi.org/10.3390/s22145160

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop