Next Article in Journal
Fractional Hermite–Hadamard-Type Inequalities for Differentiable Preinvex Mappings and Applications to Modified Bessel and q-Digamma Functions
Next Article in Special Issue
Fokker–Planck Analysis of Superresolution Microscopy Images
Previous Article in Journal
Model-Based Assessment of Elastic Material Parameters in Rheumatic Heart Disease Patients and Healthy Subjects
Previous Article in Special Issue
Modelization of Low-Cost Maneuvers for an Areostationary Preliminary Mission Design
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Study of Metaheuristic Optimization of Convolutional Neural Networks Applied to Face Mask Classification

Tijuana Institute of Technology, TecNM, Calzada Tecnológico S/N, Fracc. Tomas Aquino, Tijuana CP 22379, BC, Mexico
*
Author to whom correspondence should be addressed.
Math. Comput. Appl. 2023, 28(6), 107; https://doi.org/10.3390/mca28060107
Submission received: 18 August 2023 / Revised: 27 October 2023 / Accepted: 27 October 2023 / Published: 1 November 2023

Abstract

:
The preventive measures taken to curb the spread of COVID-19 have emphasized the importance of wearing face masks to prevent potential infection with serious diseases during daily activities or for medical professionals working in hospitals. Due to the mandatory use of face masks, various methods employing artificial intelligence and deep learning have emerged to detect whether individuals are wearing masks. In this paper, we utilized convolutional neural networks (CNNs) to classify the use of face masks into three categories: no mask, incorrect mask, and proper mask. Establishing the appropriate CNN architecture can be a demanding task. This study compares four swarm intelligent metaheuristics: particle swarm optimization (PSO), grey wolf optimizer (GWO), bat algorithm (BA), and whale optimization algorithm (WOA). The CNN architecture design involves determining the essential hyperparameters of the CNNs. The results indicate the effectiveness of the PSO and BA in achieving an accuracy of 100% when using 10% of the images for testing. Meanwhile, when 90% of the images were used for testing, the results were as follows: PSO 97.15%, WOA 97.14%, BA 97.23%, and GWO 97.18%. These statistically significant differences demonstrate that the BA allows better results than the other metaheuristics analyzed in this study.

1. Introduction

The COVID-19 pandemic has shown the importance of using face masks, avoiding the spread of the virus, and preventing the infection of millions of people [1,2]. However, it is important to mention that various studies on its use were performed several years before the COVID-19 pandemic, where the importance and efficacy of its use to prevent other respiratory infections were demonstrated [3,4]. Two of the most widely used subsets of artificial intelligence related to face masks are deep learning (DL) and machine learning (ML). Different works on the detection of the facial mask using pre-trained models of convolutional neural networks can be found in [5,6,7], which allowed us to observe the potential of this technique in the detection and classification of facial masks [8,9,10]. In Ref. [11], the authors studied the architectures of different pre-trained models such as EfficientNet, InceptionV3, MobileNetV1, MobileNetV2, ResNet-101, ResNet-50, VGG16, and VGG19. Based on their study, they proposed a model for face mask detection based on MobileNetV2, applying data augmentation techniques to increase the number of images for the training phase. In Ref. [12], an application for mobile devices was developed to identify face masks using the Google Cloud ML API, while analyzing the progress of cloud technology and the benefits of machine learning. In Ref. [10], the authors developed five ML models for face mask classification. The developed models were Naïve Bayes (NB), Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), and K-Nearest Neighbors (KNN). The test of the models was performed using 1222 images, where the results demonstrated the effectiveness of the DT over the other models. The use of neural networks is related to metaheuristics, which are utilized to find the optimal architectures that improve the results depending on the application for which the network is used [13]. Metaheuristics are a great option for finding optimal parameters in applications in different areas. These algorithms have been classified according to their inspiration: based on evolutionary algorithms, physics-based algorithms, and algorithms based on swarm intelligence [14,15,16]. Nature-inspired algorithms are mainly inspired by collective behavior, where the main characteristics of a particular species are analyzed and represented in a computational way to be used in solving complex problems in the search for optimal solutions [17,18]. In recent works, comparisons have been made between metaheuristics to compare the performances applied to find CNN hyperparameters. Some of these metaheuristics are grey wolf optimizer (GWO), whale optimization algorithm (WOA), salp swarm algorithm (SSA), sine cosine algorithm (SCA), multiverse optimizer (MVO), particle swarm optimization (PSO), moth flame optimization (MFO), and bat algorithm (BA), to mention a few. The authors have concluded the advantages of combining convolutional neural networks and metaheuristics for the search of hyperparameters [19,20,21]. These techniques have been combined to solve applications related to pattern recognition [19,22,23], image classifications [18,24], and medical diagnosis [21,25,26], among other applications.
In this work, convolutional neural network hyperparameters are optimized by different nature-inspired algorithms [27,28,29]. The optimized hyperparameters are the number of convolutional layers, filters, fully connected layers, neurons, batch size, and epochs. The contribution of this work includes the optimal design of the convolutional neural network architectures to increase classification accuracy and its application to face mask classification: no mask, incorrect mask, and mask. Recent works applied to face mask classification based their model architectures on pre-trained models, which does not guarantee optimal architecture. As a novelty, this paper proposed optimizing CNN architectures instead of basing them on other architectures. The optimal hyperparameters are found using four metaheuristics used in recent works to make a statistical comparison and analysis, providing better accuracy for face mask classification.
This paper is presented as follows. In Section 2, the metaheuristics applied in this work are presented in a succinct manner. Section 3 shows the optimization proposed for the convolutional neural networks. The results obtained by each swarm intelligence metaheuristic are shown in Section 4. The statistical test results are shown in Section 5. The conclusions are presented in Section 6.

2. Background

The optimal design of architectures and models has allowed the realization of important practical applications. In Ref. [30], optimal convolutional neural network architectures was designed to identify various types of damage on reinforced concrete (RC) to avoid further structure deterioration. The results achieved show good accuracy of six types of damage. The design of convolutional neural network architectures using a particle swarm optimization algorithm was proposed and applied to sign language recognition using three study cases of sign language databases: the Mexican Sign Language alphabet, the American Sign Language MNIST, and the American Sign Language alphabet [31]. In Ref. [32], the authors proposed face detection and face classification by developing adaptive sailfish moth flame optimization (ASMFO) to the parameter optimization using a deep learning approach. In Ref. [19], the authors analyzed the importance of the CNN hyperparameters, such as filters, kernel, epoch, batch size, and pooling size of the convolutional neural networks applied to classify human movements. They compared seven metaheuristic algorithms: GWO, WOA, SSA, SCA, MVO, PSO, and MFO, concluding the advantages of the metaheuristics to optimize the hyperparameters of CNNs. The results led the authors to the conclusion that the implementation of GWO achieved higher accuracy than the other metaheuristics. In Ref. [20], the authors proposed a PSO to determine optimal hyperparameters of convolutional neural networks. They used the simplest CNN model as a base: LeNet. Their results achieved better results when the PSO designed the CNN architectures. The results achieved by their study were obtained using MNIST, Fashion-MNIST, and CIFAR-10 datasets.
In previous works [33,34], nature-inspired algorithms have optimized modular neural network architectures applied to human recognition using different biometric measures. In those works, comparisons using genetic algorithms (GAs) and swarm intelligence algorithms were performed, and significant evidence of the advantage of the swarm intelligence algorithms was proven. More recently, in [22], and based on the advantages offered by the swarm intelligence algorithms, the architecture of convolutional neural networks was optimized and applied to face recognition. In this work, algorithms such as particle swarm optimization and grey wolf optimizer offer advantages when designing convolutional neural network architectures. It is important to mention that the databases used for this work were small, with 400 and 165 images. In Ref. [35], the non-optimized design of convolutional neural network architectures applied to the facial mask classification was performed, and the best architecture was implemented in a real-time system using a Raspberry Pi 4 in combination with a camera to obtain the image in real time. The Raspberry Pi 4 sends a signal through its GPIO Board, and a result is provided by lighting an LED. If the mask is correctly used, the green LED is turned on. If the mask is incorrectly used, the yellow LED is turned on, and if a mask is not used, the red LED is turned on.

3. Intelligence Techniques

This section shows a description of the intelligence techniques utilized in this work.

3.1. Convolutional Neural Networks

Artificial neural networks (ANNs) are mainly based on the behavior of the human nervous system and its way of processing information. An artificial neural network is a type of distributive processor made up of simple processing units known as neurons, simulating two main aspects of the human brain: it acquires knowledge of its environment through a learning process and the use of synaptic weights to store the required knowledge [36,37]. Learning methods are categorized into supervised, semi-supervised, and unsupervised learning. Among the main properties that can be found in ANNs that make them one of the main techniques used in artificial intelligence, we can find their capacity for generalization, adaptation, learning, and parallelism [38,39]. Convolutional neural networks (CNNs) are an improvement of ANNs with some characteristics that make them powerful in applications where images are used. This type of network consists of other layers in addition to those already existing in conventional neural networks: the convolutional and the pooling layers. One of the advantages provided by this type of network is the extraction of features from the given images before proceeding to the learning phase, which makes it possible to reduce the amount of information that must be learned by the ANN [38]. In the convolutional layers (CLs), the inputs are multiplied by a filter with the size m × n. Each layer contains a height, width, and depth. When talking about depth concerning the layer, it refers to the number of channels (primary colors) that contain the input images [40]. The most used grouping layers with the maximum, average, and minimum are responsible for grouping the feature map produced in the convolution layer, thus reducing the amount of information that will pass to the fully connected layers [23,41]. In Figure 1, a representation of a convolutional neural network is shown.

3.2. Nature-Inspired Algorithms

The nature-inspired algorithms used in this study are described below.

3.2.1. Particle Swarm Optimization

In Ref. [42], the particle swarm optimization (PSO) based on the fish or bird social performance was proposed. A set of particles is known as a swarm, and each particle is a solution [43]. A particle defines their next position by Equation (1).
x i d t + 1 = x i d t + v i d t + 1
where x i d t indicates at time t, in the dimension d, the actual position of the particle i. A velocity v i t + 1 is designated to establish the next position. In Ref. [44], this algorithm was enhanced by adding the parameter: inertia weight (w). The particle velocity is defined by Equation (2).
v i d t + 1 = w × v i d t + c 1 × r 1 d t × y i d t x i d t + c 2 × r 2 d t × y ^ d t x i d t
where r 1 and r 2 are random values in [0, 1]. The best position of a particle i in dimension d is connoted by y i d t ; the best position of the swarm in d dimension is denoted by y ^ d t .   c 1 and c 2 are the cognitive and social components.
w has a decreased value during the algorithm execution to allow exploitation and exploration. The linear decrease in the inertia weight applied in this work is given by Equation (3).
w t = w s w e × ( t m a x t ) t m a x + w e
where t m a x denotes the maximum number of time steps, and w s and w e are the initial and final values of the inertia weight, respectively. The recommended values are w s = 0.9 and w e = 0.4 [45].

3.2.2. Grey Wolf Optimizer

In Ref. [46], the grey wolf optimizer (GWO) was proposed. This metaheuristic uses a dominant hierarchy applied by the wolves in hunting as inspiration. This dominant hierarchy is shown in Figure 2, where leaders known as alphas are at the top of the pyramid, and they make the main hunting and sleeping decisions. The betas are subaltern wolves that help the alpha wolves in making decisions. A delta wolf does not belong to any level already mentioned and can dominate only the lowest level. The delta wolves have different roles as scouts, sentinels, elders, hunters, and caretakers. The wolves in the lowest level are known as the omegas. They are always submitted by the wolves that are in the superior hierarchies [47,48].
The description of the principles used to define this algorithm and its mathematical representation is described below:
  • Social hierarchy: The three best solutions are alpha (α), beta (β), and delta (δ). The wolves belonging to the lowest level are the omegas (ω).
  • Encircling prey: The process of prey encircling during hunting are represented by Equations (4) and (5).
D = |   C × X p   t X   t
X t + 1 = X p t A × D
where X denotes the agent position in the t iteration, and X p represents the position of the prey. The coefficient vectors are A and C . Equations (6) and (7) are used to determine their values.
A = 2 a × r 1 a  
C = 2 × r 2
where r 1 and r 2 represent vectors with random values in [0, 1]. During the algorithm execution, the vector a has linear decreasing values in [2, 0] given by Equation (8) [49].
a t = 2 2 × t t m a x
where t denotes the current iteration, and t m a x denotes the maximum number of iterations.
  • Hunting: The first three levels in the dominant hierarchy know the prey position. With their positions, the wolves belonging to the lowest level (omega) can update their position using Equations (9)–(11).
D α =   C 1 × X α X ,           D β =   C 2 × X β X ,         D δ =   C 3 × X δ X
X 1 =   X α A 1 × D α ,         X 2 =   X β A 2 × D β ,           X 3 =   X δ A 3 × D δ
X t + 1 = X 1 + X 2 + X 3 3
  • Attacking prey: The process is also known as exploitation, where the current position of an agent and the prey allows it to establish the next position of the agent. This position is calculated using a and vector A with random values in an interval [−2 a ,   2 a ].
  • Search for prey: The process is also known as exploration, where vector C is used with values in [0, 2] to provide diversity to the population and avoid local optimal.

3.2.3. Whale Optimization Algorithm

In Ref. [50], the whale optimization algorithm (WOA) was proposed. This algorithm uses as inspiration the hunting method applied by the whales. These marine mammals usually live in groups and are considered killers and predators [51]. One of the main characteristics shared with the grey wolf optimizer is the process of encircling prey, present also in WOA. The description of the processes used to define this algorithm based on humpback whales and its mathematical representation is described below:
  • Encircling prey: The whales encircle the prey because they know its position. The whale closest to the prey becomes the best solution. Equations (3) and (4) allow the update of the position of the rest of the agents.
  • Bubble-net attacking method: This process is also known as exploitation and is very similar to the one in the GWO, where the distance between the agent and the prey is determined. The process can be accomplished using two approaches:
    • Mechanism of shrinking encircling: In Equation (5), the values of a decrease every iteration, and an interval [ a ,   a ] is used to generate random values for the vector A .
    • Spiral updating position: The helix-shaped movement of whales between the whale and prey position is mimicked by Equation (12).
X t + 1 = D × e b l × cos 2 π l + X p t
where the distance between prey and whale is connoted by D , and b is a constant that represents the shape of the logarithmic spiral. A random value in an interval [−1, 1] is represented by l.
  • Search for prey: This process is also known as exploration, where the whales seek randomly based on the position of others. To force the exploration, the A vector has numbers less than −1 and greater than 1. The process is defined by Equations (13) and (14).
D =   C × X r a n d X
X t + 1 = X r a n d A × D
where a random whale of the current iteration is represented by X r a n d . This random whale or the best solution found is utilized to help the other whales update their position.

3.2.4. Bat Algorithm

The bat algorithm (BA) [52] is based on their echolocation behavior due to the ability they have to identify their prey even in darkness. There are different types of bats depending on their size. Microbats have the characteristic of using a type of sonar known as echolocation, allowing them to detect prey and avoid obstacles. The bats make a loud sound pulse and listen for the echo that is reflected off of nearby objects. The rate of pulse is established in an interval [0, 1] [53].
The author established some important rules to delimit the behavior and knowledge that the bats can have:
  • Echolocation is used for all the bats to sense distance, and they know the difference between the prey and other kind of elements.
  • To search for prey, each bat flies randomly in a position x i with a velocity v i . This task is performed by changing loudness A and wavelength λ . Depending on the closeness of its objective, the bat regulates the wavelength of its emitted pulses and regulate the rate of pulse emission r [ 0,1 ] .
  • The loudness is assumed to be a large value positive number A to a minimum constant value A m i n .
To define the update of position and velocities, the next equations are given by Equations (15)–(17).
f i = f m i n + f m a x + f m i n × β
v i t = v i t 1 + x i t 1 x × f i
x i t = v i t 1 + v i t
where x i t and v i t represent the new position and velocity, respectively, at time step t . A vector with random values in [0, 1] is represented by β . The current global best solution is denoted by x . For the local search, the best solutions are used to select one of them and locally generate a new solution using a random walk given by Equation (18) [54].
x n e w = x o l d + ε × A ( t )
where ε represents a random value in an interval [−1, 1], and the average loudness of all the bats at time step t is represented by A ( t ) .

4. Proposed Method

The proposed optimization is applied to face mask classification (no mask, incorrect mask, and mask). To perform this task, the method combines CNNs and optimization algorithms. The metaheuristics allow the optimal design of CNN architectures to be found. The optimization algorithm designs the CNN architectures, seeking their number of convolutional layers, filters, fully connected layers, neurons, batch size, and epochs. Each CL is followed by a max-pooling layer with a pool size of 3 × 3 to reduce image size. Figure 3 shows an example of the CNN architecture applied to face mask classification. As input to the convolutional neural network, images of people wearing (correctly and incorrectly) or not using face masks are used for the training phase of the convolutional neural network. The first layers of the convolutional neural network (convolutional and pooling) will extract features and reduce the image so that the fully connected layers learn the most relevant information. As output, when an image is simulated, a classification will be obtained (no mask, incorrect mask, and mask). A correct classification will depend on correct learning and the convolutional neural network architecture. For this reason, an optimization algorithm is an excellent option for designing the architecture because it allows a specific model to be applied to a particular application.

4.1. Description of the Optimization

The parameters used to execute any optimization algorithm have great importance because these depend on its performance. For each optimization algorithm, 10 solutions (particles, bats, or search agents) and 10 iterations are used. The configuration of the optimization algorithms used in this work is presented in more detail in Table 1. The parameters presented are based on previous works [22,33,55].
Each solution seeks to minimize the face mask classification error. In this work, the accuracy equation is used and given by Equation (19).
A c c u r a c c y = T P + T N T P + F P + T N + F N
where TP, TN, FP, and FN mean True Positive, True Negative, False Positive, and False Negative, respectively. The objective function used in this work is expressed by Equation (20).
f = 1 T P + T N T P + F P + T N + F N
The search space used for each solution (particle or agent) is determined by the minimum and maximum ranges shown in Table 2. These ranges are established based on previous works [18,22]. The convolutional neural networks use the Adaptive Moment Estimation (Adam) as a leaning algorithm and the rectified linear activation function (ReLU) as an activation function. The batch size is determined using a range from 1 up to 5, which means 8, 16, 32, 64, or 128.
Each particle or agent represents a solution, where each solution has 14 dimensions, which allow the creation of a CNN. In Figure 4, the dimensions of the solution are shown. The first four dimensions allow the determination of the number of convolutional layers, epoch, batch size, and the number of fully connected layers. Meanwhile, the rest of the dimensions allow us to determine the number of neurons and filters.
All the metaheuristics have, as a stopping criterion, 10 iterations or when the best solution has a cost equal to zero. The Keras Python package based on TensorFlow was used to implement the optimization algorithms and to build and train the CNN models.

4.2. Database

To perform the face mask classification, the convolution neural networks are trained, validated, and tested using images of three classes (no mask, incorrect mask, and mask). The first two classes are obtained from the MaskedFace-Net dataset [56], and the no mask class is obtained from the Flickr-Faces-HQ Dataset (FFHQ) [57]. The MaskedFace-Net dataset consists of 137,016 images, and it is based on the Flickr-Faces-HQ (FFHQ) dataset. In this work, 3000 images were used, where each class contains 1000 images of the dataset. In Figure 5, a sample of the dataset is shown. The images used in this work were separated into training, validation, and testing. To help prevent bias in our models, when the images are split into sets, stratified sampling is utilized to guarantee a consistent distribution. Stratified sampling is a functionality provided by the Keras Python package.

4.3. Preprocessing

The original images have a resolution of 1024 × 1024 pixels. The region of interest (ROI) for this work is the face region, and it is automatically found using the Caffe model. The Caffe model was developed by the Berkeley Vision and Learning Center (BVLC). This model was trained to perform object detection and classification [58]. In Figure 6, an example of the face detection is shown.
When the face region is detected, the image is resized to 100 × 100 pixels. Once the image is resized, an RGB subtraction technique is implemented to the ROI in order to help counteract slight variations [59]. In Figure 7, an example of the RGB subtraction technique is shown.
The proposed method is shown in Figure 8, which begins with the input images that go through preprocessing. The database is partitioned into three sets (training, validation, and testing), and the optimized CNN architecture is obtained with the metaheuristic.

5. Experimental Results

The database previously described is used to prove the proposed hyperparameters optimization. As previously mentioned, 3000 images were used to train, validate and test each convolutional neural network. In this work, 20 runs were performed using 10, 20, 30, 40, 50, 70, 80, and 90 percent of the images for testing, leaving the rest for training and validation. These experiments are performed with all the previously mentioned metaheuristics.

5.1. PSO Results

The best architectures achieved by the PSO with different percentages of images for the testing phase are summarized in Table 3. The best results are obtained with 10% and 20% of images for the testing phase, where an accuracy of 100% is achieved (marked with bold text in Table 3). We can define the best architecture as the one that uses less information for the training phase, which would be when 20% is used for the testing phase. This CNN model is structured as follows: four convolutional layers with 16, 16, 28, and 23 filters, with a size of 3 × 3. This architecture uses four FCLs with 150, 10, 117, and 19 neurons and a batch size of 8 with 12 epochs.
The results achieved by the PSO are shown in Table 4. The results illustrate how the accuracy (best and average) decreases as the percentage of images for the testing phase increases, and this occurs because the CNN is trained with less information.

5.2. WOA Results

In Table 5, the best architectures achieved by the WOA with different percentages of images for the testing phase are shown. The best results are also obtained with 10% and 20% of images for the testing phase, where an accuracy of 100% is achieved (marked with bold text in Table 5). The best architecture can be defined as the one that uses 20% for the testing phase. This CNN model is structured as five CLs with 16, 16, 32, 32, and 64 filters, with a size of 3 × 3 with five FCLs with 150, 88, 150, 100, and 50 neurons and a batch size of 32 with 20 epochs.
The results achieved by the WOA is shown in Table 6. The results show how the accuracy (best and average) also decreases as the percentage of images for the testing phase increases, except the best result using 40% of the images in the testing phase, which is superior to the best value obtained using 30%.

5.3. BA Results

The best architectures achieved by the BA with different percentages of images for the testing phase are shown in Table 7. The best result is obtained with only 10% of images for the testing phase, where an accuracy of 100% is achieved (marked with bold text in Table 7). This CNN model is structured as follows: three CLs with 11, 10, and 28 filters, with a size of 3 × 3, three FCLs with 121, 61, and 63 neurons, and a batch size of 16 with 14 epochs. This architecture uses less convolutional and fully connected layers than the previous ones, which also obtained 100% accuracy.
Table 8 shows the results achieved by the BA. For this metaheuristic, the accuracy (best and average) also decreases as the percentage of images for the testing phase increases, except the best result using 50% of the images in the testing phase, which is superior to the best value obtained using 40%.

5.4. GWO Results

Table 9 shows the best architectures achieved using the GWO with different percentages of images for the testing phase. The best results are obtained with 10% and 20% of images for the testing phase as PSO and WOA, where an accuracy of 100% is achieved (marked with bold text in Table 9). The best architecture can be defined as the one that uses 20% for the testing phase. This CNN model is structured in the following way: four convolutional layers with 10, 8, 23, and 25 filters, with a size of 3 × 3 with three FCLs with 42, 139, and 32 neuron and a batch size of 8 with 20 epochs.
Table 10 shows the results obtained by the GWO. The accuracy decreases as with the other metaheuristics, but when 30% and 50% for the testing phase are used, the same result (the best value) is obtained. It is important to mention that when using 40%, the accuracy is better (the best value).

5.5. Comparison of Results

In Table 3, Table 5, Table 7 and Table 9, the best architectures generated by each metaheuristic are presented, where it can be seen how the architectures can vary and still provide good results without using architectures as complex as those of the pre-trained models.
In Figure 9, the accuracy values (best, average, and worst) shown in Table 4, Table 6, Table 8 and Table 10 are graphically shown. We can see that the PSO (Figure 9a) and BA (Figure 9c) always achieve an accuracy of 100% when 10% of the images are used for the testing phase (90% for training and validation). Meanwhile, the WOA and GWO only achieved the same value in some experiments using the same percentage of images. Using 50% of the percentage of images, we can see how the BA and PSO have very parallel values, which indicates that there is not much difference between their values (best, average, and worst), which could indicate greater stability between the results obtained in their experiments.
The average convergence during the learning phase of the 20 runs for each percentage of images (from 10 up to 90) obtained with each metaheuristic is depicted in Figure 10. It can be observed that when different percentages of images are used for the testing phase, the behavior of the PSO and BA is very similar. Even with only 10% of the images used for testing, both the PSO and BA achieve an error of 0 by iterations 6 and 4, respectively. Meanwhile, WOA and GWO exhibit similar behavior when 60%, 70%, 80%, and 90% of the images are used for the testing phase.
The accuracy and loss curves with their respective validation of the best models are depicted in Figure 11. These models achieved an accuracy of 100%. The figure shows how the accuracy and loss have similar behavior to their validation.
Table 11 shows the averages (accuracy) obtained by each optimization algorithm. As results show, the average decreases when the percentage of images for testing increases, which means the CNN has less information to learn. Only two metaheuristics can achieve 100% accuracy: PSO and BA. Figure 12 shows graphically the accuracy achieved by the metaheuristics.
The errors achieved by each metaheuristic are shown in Table 12. These results are utilized to perform statistical comparisons in the next section. These errors are graphically shown in Figure 13.
Figure 12 and Figure 13 graphically show the results obtained in this study. It can be seen that when a percentage between 10% and 50% is used for the testing phase, the PSO, WOA, and BA have similar good behavior. Meanwhile, when the percentage increases, it can be observed that the BA has a better accuracy, which means less error. In Table 13, the results achieved with the best average of accuracy are shown using other metrics (Recall, Precision, and F1 Score). The results show that the BA achieved better results in the other metrics, proving the effectiveness in metrics such as the F1 Score, where a combination of Recall and Precision is performed.

6. Statistical Comparison

This section shows statistical comparisons where the averages (errors) achieved by each optimization algorithm are used. In this work, the Wilcoxon signed-rank tests are utilized, where the value of α depends on the statistical significance. Table 14 shows the critical values with different statistical significance levels. A significance level of 0.10 is used in this work.
Table 15 shows the results of the statistical tests performed among all the metaheuristics. The null hypothesis assumes that means are equal, which contradicts the alternative hypothesis. The null hypothesis can be rejected if the column “W” value is equal to or smaller than the “W0” based on the critical value with a 0.10 significance level. All possible comparisons were performed among the four metaheuristics studied in this work. The results exhibit a significant difference between the PSO and GWO. Meanwhile, the BA achieves significant differences against the other metaheuristics, allowing a better face mask classification.
The results obtained with the method applying the bat algorithm allowed us to obtain better results, especially when less percentage is utilized for the training phase of the CNNs applied to face mask classification.

7. Conclusions

In this work, four swarm intelligence metaheuristics were applied to perform a comparison. A face mask database is used as a training, validation, and testing set to prove the proposed CNN design. This database has three classes: no mask, incorrect mask, and mask. The metaheuristics applied to CNN architecture design were PSO, WOA, BA, and GWO. These algorithms were implemented to CNN optimization applied to face mask multiclass classification, where hyperparameters of CNN were sought: the number of convolutional layers, filters, number of fully connected layers, neurons, batch size, and epoch. The results showed that some average convergences of the metaheuristics have a similar behavior when different percentages of images for the testing phase are utilized. The PSO and BA achieved an average of 100% accuracy when 10% of the images for the testing phase were used (leaving 90% for training and validation), but the BA converged faster than the PSO. The Wilcoxon signed-rank tests are utilized to compare results, and there is a statistical difference when the PSO and GWO are compared. However, when comparing the BA against PSO, WOA, and GWO, there is a statistical difference, which indicates that the BA allows for achieving better results than the other metaheuristics analyzed in this study when hyperparameters of convolutional neural networks are searched for the face mask classification. Results achieved in previous works and the results obtained in this work show that the performance of each optimization algorithm will depend on its application. In this work, only 3000 images were used, and different percentages of images were used for each phase to find optimal architectures with fewer images performing comparisons among swarm intelligence algorithms. The real implementation implies that optimized models have learned enough with the idea of not invading privacy and not having to train the models with specific persons. The optimized architectures could perform a correct face mask classification independently whether images of a person were used or not to train the model. The comparison performed in this work will allow us, as future work, to select those optimized architectures with a better percentage of accuracy and continue with the implementation in a real-time system. Although metaheuristics allow for optimal architectures with high accuracy, several limitations must be addressed in future works, such as the use of novel types of face masks not considered in this work, which would lead us to the need to evaluate their behavior. The dataset used to train and evaluate the architectures uses different face positions. However, it would be important to work with images with different kinds of illumination, especially for future work on implementing these architectures in real systems. Also, in future works, the comparison of these metaheuristics will be implemented by applying them to other intelligent techniques, such as fuzzy logic for parameter adjustment or fuzzy control.

Author Contributions

Methodology and validation, P.M.; Software, validation, writing, D.S.; Conceptualization, creation on main idea, writing—review and editing, O.C.; formal analysis, M.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank TecNM and Conacyt for their support during the realization of this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Eikenberry, S.; Mancuso, M.; Iboi, E.; Phan, T.; Eikenberry, K.; Kuang, Y.; Kostelich, E.; Gumel, A. To mask or not to mask: Modeling the potential for face mask use by the general public to curtail the COVID-19 pandemic. Infect. Dis. Model. 2020, 5, 293–308. [Google Scholar] [CrossRef] [PubMed]
  2. Garcia Godoy, L.; Jones, A.; Anderson, T.; Fisher, C.; Seeley, K.; Beeson, E.; Zane, H.; Peterson, J.; Sullivan, P. Facial protection for healthcare workers during pandemics: A scoping review. BMJ Glob. Health 2020, 5, e002553. [Google Scholar] [CrossRef] [PubMed]
  3. MacIntyre, C.; Cauchemez, S.; Dwyer, D.; Seale, H.; Cheung, P.; Browne, G.; Fasher, M.; Wood, J.; Gao, Z.; Booy, R.; et al. Face Mask Use and Control of Respiratory Virus Transmission in Households. Emerg. Infect. Dis. 2009, 15, 233–241. [Google Scholar] [CrossRef] [PubMed]
  4. MacIntyre, C.; Chughtai, A.; Rahman, B.; Peng, Y.; Zhang, Y.; Seale, H.; Wang, X.; Wang, Q. The efficacy of medical masks and respirators against respiratory infection in healthcare workers. Influenza Other Respir. Viruses 2017, 11, 511–517. [Google Scholar] [CrossRef] [PubMed]
  5. Pham-Hoang-Nam, A.; Le-Thi-Tuong, V.; Phung-Khanh, L.; Ly-Tu, N. Densely Populated Regions Face Masks Localization and Classification Using Deep Learning Models. In Proceedings of the Sixth International Conference on Research in Intelligent and Computing, Thủ Dầu Một, Vietnam, 3–4 June 2021. [Google Scholar]
  6. Sethi, S.; Kathuria, M.; Kaushik, T. Face mask detection using deep learning: An approach to reduce risk of Coronavirus spread. J. Biomed. Inform. 2021, 120, 103848. [Google Scholar] [CrossRef]
  7. Yu, J.; Zhang, W. Face Mask Wearing Detection Algorithm Based on Improved YOLO-v4. Sensors 2021, 21, 3263. [Google Scholar] [CrossRef]
  8. Mar-Cupido, R.; Garcia, V.; Rivera, G.; Sánchez, J. Deep transfer learning for the recognition of types of face masks as a core measure to prevent the transmission of COVID-19. Appl. Soft Comput. 2022, 125, 109207. [Google Scholar] [CrossRef]
  9. Umer, M.; Sadiq, S.; Alhebshi, R.; Alsubai, S.; Hejaili, A.; Eshmawi, A.; Nappi, M.; Ashraf, I. Face mask detection using deep convolutional neural network and multi-stage image processing. Image Vis. Comput. 2023, 133, 104657. [Google Scholar] [CrossRef]
  10. Ramakrishnan, K.; Balakrishnan, V.; Wong, H.; Tay, S.; Soo, K.; Kiew, W. Face Mask Wearing Classification Using Machine Learning. Eng. Proc. 2023, 41, 13. [Google Scholar]
  11. Habib, S.; Alsanea, M.; Aloraini, M.; Al-Rawashdeh, H.; Islam, M.; Khan, S. An Efficient and Effective Deep Learning-Based Model for Real-Time Face Mask Detection. Sensors 2022, 22, 2602. [Google Scholar] [CrossRef]
  12. Wakchaure, A.; Kanawade, P.; Jawale, M.; William, P.; Pawar, A. Face Mask Detection in Realtime Environment using Machine Learning based Google Cloud. In Proceedings of the International Conference on Applied Artificial Intelligence and Computing, Salem, India, 9–11 May 2022. [Google Scholar]
  13. Mirjalili, S. Evolutionary Algorithms and Neural Networks: Theory and Applications, 1st ed.; Springer: London, UK, 2019. [Google Scholar]
  14. Du, K.; Swamy, M. Search and Optimization by Metaheuristics: Techniques and Algorithms Inspired by Nature, 1st ed.; Birkhäuser Cham: Berlín, Germany, 2018. [Google Scholar]
  15. Hassanien, A.; Emary, E. Swarm Intelligence: Principles, Advances and Applications, 1st ed.; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  16. Iba, H. AI and SWARM: Evolutionary Approach to Emergent Intelligence, 1st ed.; CRC Press: Boca Raton, FL, USA, 2019. [Google Scholar]
  17. Poma, Y.; Melin, P.; Gonzalez, C.; Martinez, G. Optimization of convolutional neural networks using the fuzzy gravitational search algorithm. J. Autom. Mob. Robot. Intell. Syst. 2020, 14, 109–120. [Google Scholar] [CrossRef]
  18. Yang, X. Nature-Inspired Computation and Swarm Intelligence: Algorithms, Theory and Applications, 1st ed.; Academic Press: Cambridge, MA, USA, 2020. [Google Scholar]
  19. Raziani, S.; Azimbagirad, M. Deep CNN hyperparameter optimization algorithms for sensor-based human activity recognition. Neurosci. Inform. 2022, 2, 100078. [Google Scholar] [CrossRef]
  20. Yeh, W.; Lin, Y.; Liang, Y.; Lai, C.; Huang, C. Simplified swarm optimization for hyperparameters of convolutional. Comput. Ind. Eng. 2023, 177, 109076. [Google Scholar] [CrossRef]
  21. Chawla, R.; Beram, S.; Murthy, C.; Thiruvenkadam, T.; Bhavani, N.; Saravanakumar, R.; Sathishkumar, P. Brain tumor recognition using an integrated bat algorithm with a convolutional neural network approach. Meas. Sens. 2022, 24, 100426. [Google Scholar] [CrossRef]
  22. Melin, P.; Sánchez, D.; Castillo, O. Comparison of optimization algorithms based on swarm intelligence applied to convolutional neural networks for face recognition. Int. J. Hybrid Intell. Syst. 2022, 18, 161–171. [Google Scholar] [CrossRef]
  23. Melin, P.; Sánchez, D.; Pulido, M.; Castillo, O. Convolutional Neural Network Design using a Particle Swarm Optimization for Face Recognition. In Proceedings of the International Conference on Hybrid Intelligent Systems, Online, 13–15 December 2021. [Google Scholar]
  24. Fernandes Junior, F.; Yen, G. Particle swarm optimization of deep neural networks architectures for image classification. Swarm Evol. Comput. 2019, 49, 62–74. [Google Scholar] [CrossRef]
  25. Bashkandi, A.; Sadoughi, K.; Aflaki, F.; Alkhazaleh, H.; Mohammadi, H.; Jimenez, G. Combination of political optimizer, particle swarm optimizer, and convolutional neural network for brain tumor detection. Biomed. Signal Process. Control 2023, 81, 104434. [Google Scholar] [CrossRef]
  26. Murugan, R.; Goel, T.; Mirjalili, S.; Chakrabartty, D. WOANet: Whale optimized deep neural network for the classification of COVID-19 from radiography images. Biocybern. Biomed. Eng. 2021, 41, 1702–1708. [Google Scholar] [CrossRef]
  27. Knypiński, Ł. Constrained optimization of line-start PM motor based on the gray wolf optimizer. Maint. Eng. 2021, 23, 1–10. [Google Scholar] [CrossRef]
  28. Nazri, E.; Murairwa, S. Classification of heuristic techniques for performance comparisons. In Proceedings of the International Conference on Mathematics, Statistics, and Their Applications, Banda Aceh, Indonesia, 4–6 October 2016. [Google Scholar]
  29. Kumar, A.; Bawa, S. A comparative review of meta-heuristic approaches to optimize the SLA violation costs for dynamic execution of cloud services. Soft Comput. 2020, 24, 3909–3922. [Google Scholar] [CrossRef]
  30. Fan, C.; Chung, Y. Design and Optimization of CNN Architecture to Identify the Types of Damage Imagery. Mathematics 2022, 10, 3483. [Google Scholar] [CrossRef]
  31. Fregoso, J.; Gonzalez, C.; Martinez, G. Optimization of Convolutional Neural Networks Architectures Using PSO for Sign Language Recognition. Axioms 2021, 10, 139. [Google Scholar] [CrossRef]
  32. Shaban Naseri, R.; Kurnaz, A.; Farhan, H. Optimized face detector-based intelligent face mask detection model in IoT using deep learning approach. Appl. Soft Comput. 2023, 134, 109933. [Google Scholar] [CrossRef]
  33. Sánchez, D.; Melin, P.; Castillo, O. A Grey Wolf Optimizer for Modular Granular Neural Networks for Human Recognition. Comput. Intell. Neurosci. 2017, 2017, 4180510. [Google Scholar] [CrossRef]
  34. Sánchez, D.; Melin, P.; Castillo, O. Optimization of modular granular neural networks using a firefly algorithm for human recognition. Eng. Appl. Artif. Intell. 2017, 64, 172–186. [Google Scholar] [CrossRef]
  35. Campos, A.; Melin, P.; Sánchez, D. Multiclass Mask Classification with a New Convolutional Neural Model and Its Real-Time Implementation. Life 2023, 13, 368. [Google Scholar] [CrossRef]
  36. Haykin, S. Neural Networks: A Comprehensive Foundation, 1st ed.; Macmillan: London, UK, 1994. [Google Scholar]
  37. Nunes Da Silva, I.; Hernane Spatti, D.; Flauzino, A.; Bartocci Liboni, L.; Dos Reis Alves, S. Artificial Neural Networks: A Practical Course, 1st ed.; Springer: London, UK, 2018. [Google Scholar]
  38. Aggarwal, C. Neural Networks and Deep Learning: A Textbook, 1st ed.; Springer: London, UK, 2018. [Google Scholar]
  39. Singh, M.; Singh, G. Two phase learning technique in modular neural network for pattern classification of handwritten Hindi alphabets. Mach. Learn. Appl. 2021, 6, 100174. [Google Scholar] [CrossRef]
  40. Koonce, B. Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization, 1st ed.; Apress: New York, NY, USA, 2021. [Google Scholar]
  41. Ozturk, S. Convolutional Neural Networks for Medical Image Processing Applications, 1st ed.; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar]
  42. Eberhart, R.; Kennedy, J. A New Optimizer using Particle Swarm. In Proceedings of the International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995. [Google Scholar]
  43. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the IEEE International Joint Conference on Neuronal Networks, Perth, WA, Australia, 27 November–1 December 1995. [Google Scholar]
  44. Eberhart, R.; Shi, Y. Comparing Inertia Weights and Constriction Factors in Particle Swarm Optimization. In Proceedings of the IEEE Congress on Evolutionary Computation, La Jolla, CA, USA, 16–19 July 2000. [Google Scholar]
  45. Xin, J.; Chen, G.; Hai, Y. A Particle Swarm Optimizer with Multi-stage Linearly-Decreasing Inertia Weight. In Proceedings of the International Joint Conference on Computational Sciences and Optimization, Sanya, China, 24–26 April 2009. [Google Scholar]
  46. Mirjalili, S.; Mirjalili, S.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  47. Mech, L. Alpha status, dominance, and division of labor in wolf packs. Can. J. Zool. 1999, 77, 1196–1203. [Google Scholar] [CrossRef]
  48. Muro, C.; Escobedo, R.; Spector, L.; Coppinger, R. Wolf-pack (Canis lupus) hunting strategies emerge from simple rules in computational simulations. Behav. Process. 2011, 88, 192–197. [Google Scholar] [CrossRef]
  49. Long, W.; Jiao, J.; Liang, X.; Tang, M. Inspired grey wolf optimizer for solving large-scale function optimization problems. Appl. Math. Model. 2018, 60, 112–126. [Google Scholar] [CrossRef]
  50. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  51. Watkins, W.; Schevill, W. Aerial Observation of Feeding Behavior in Four Baleen Whales: Eubalaena glacialis, Balaenoptera borealis, Megaptera novaeangliae, and Balaenoptera physalus. J. Mammal. 1979, 60, 155–163. [Google Scholar] [CrossRef]
  52. Yang, X. A New Metaheuristic Bat-Inspired Algorithm. In Studies in Computational Intelligence, 1st ed.; González, J.R., Pelta, D.A., Cruz, C., Terrazas, G., Krasnogor, N., Eds.; Springer: London, UK, 2010; Volume 284, pp. 65–74. [Google Scholar]
  53. Talbi, N. Design of Fuzzy Controller rule base using Bat Algorithm. Energy Procedia 2019, 162, 241–250. [Google Scholar] [CrossRef]
  54. Yang, X. Review of meta-heuristics and generalised evolutionary walk algorithm. Int. J. Bio-Inspired Comput. 2011, 3, 77–84. [Google Scholar] [CrossRef]
  55. Perez, J.; Valdez, F.; Castillo, O.; Melin, P.; Gonzalez, C.; Martinez, G. Interval type-2 fuzzy logic for dynamic parameter adaptation in the bat algorithm. Soft Comput. 2017, 21, 667–685. [Google Scholar] [CrossRef]
  56. Cabani, A.; Hammoudi, K.; Benhabiles, H.; Melkemi, M. MaskedFace-Net—A dataset of correctly/incorrectly masked face images in the context of COVID-19. Smart Health 2020, 19, 100144. [Google Scholar] [CrossRef]
  57. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  58. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding. In Proceedings of the ACM Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014. [Google Scholar]
  59. Campos, A.; Melin, P.; Sánchez, D. Convolutional neural networks for face detection and face mask multiclass classification. In Proceedings of the International Conference on Hybrid Intelligent Systems, Online, 13–15 December 2022. [Google Scholar]
Figure 1. Representation of the architecture of a convolutional neural network.
Figure 1. Representation of the architecture of a convolutional neural network.
Mca 28 00107 g001
Figure 2. The dominance hierarchy of wolves.
Figure 2. The dominance hierarchy of wolves.
Mca 28 00107 g002
Figure 3. Illustration of the CNN architecture applied to face mask classification.
Figure 3. Illustration of the CNN architecture applied to face mask classification.
Mca 28 00107 g003
Figure 4. Dimensions of the solutions to design CNN architectures. CL indicates convolutional layer, and Layer indicates fully connected layer.
Figure 4. Dimensions of the solutions to design CNN architectures. CL indicates convolutional layer, and Layer indicates fully connected layer.
Mca 28 00107 g004
Figure 5. Examples of the database with 3 classes: incorrect, mask, and no mask.
Figure 5. Examples of the database with 3 classes: incorrect, mask, and no mask.
Mca 28 00107 g005
Figure 6. Application of the Caffe Model to detect the region of interest (face detection).
Figure 6. Application of the Caffe Model to detect the region of interest (face detection).
Mca 28 00107 g006
Figure 7. Example of the RGB subtraction applied to the ROI.
Figure 7. Example of the RGB subtraction applied to the ROI.
Mca 28 00107 g007
Figure 8. The flowchart of the proposed method begins with the input images up to the optimized CNN architecture using a metaheuristic.
Figure 8. The flowchart of the proposed method begins with the input images up to the optimized CNN architecture using a metaheuristic.
Mca 28 00107 g008
Figure 9. Accuracy values achieved by: (a) PSO; (b) WOA; (c) BA; (d) GWO using different percentages of images for the testing phase.
Figure 9. Accuracy values achieved by: (a) PSO; (b) WOA; (c) BA; (d) GWO using different percentages of images for the testing phase.
Mca 28 00107 g009
Figure 10. Convergence of accuracy error for: (a) PSO; (b) WOA; (c) BA; (d) GWO using different percentages of images for the testing phase.
Figure 10. Convergence of accuracy error for: (a) PSO; (b) WOA; (c) BA; (d) GWO using different percentages of images for the testing phase.
Mca 28 00107 g010
Figure 11. The accuracy and loss curves of the best models: (a) PSO; (b) WOA; (c) BA; (d) GWO.
Figure 11. The accuracy and loss curves of the best models: (a) PSO; (b) WOA; (c) BA; (d) GWO.
Mca 28 00107 g011
Figure 12. The average accuracy obtained by the metaheuristics.
Figure 12. The average accuracy obtained by the metaheuristics.
Mca 28 00107 g012
Figure 13. The accuracy error achieved by the metaheuristics.
Figure 13. The accuracy error achieved by the metaheuristics.
Mca 28 00107 g013
Table 1. Configuration of characteristic/tuning parameters of the optimization algorithms.
Table 1. Configuration of characteristic/tuning parameters of the optimization algorithms.
PSOBATWOA and GWO
ParameterValueParameterValueParameterValue
Particles10Bats10Search Agents10
Maximum
Iterations (tmax)
10Maximum
Iterations
(tmax)
10Maximum
Iterations
(tmax)
10
C12fmin0--
C22fmax2--
w s 0.9Loudness
(A)
0.5--
w e 0.4Pulse rate
(r)
0.5--
Table 2. Definition of the search space to determine the solutions.
Table 2. Definition of the search space to determine the solutions.
HyperparameterMinimumMaximum
Convolutional layers (CLs)15
Number of filtersCL 1816
CL 2816
CL 31632
CL 41632
CL 53264
Fully connected layers (FCL) 15
Neurons10150
Epoch550
Batch Size15
Table 3. The best accuracy results and architectures obtained by the PSO. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
Table 3. The best accuracy results and architectures obtained by the PSO. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
%
Images for Testing
CLs
(Filters)
FCLs
(Neurons)
EpochBatch SizeErrorAccuracy (%)
104312320100
(12, 10, 17, 28)(65, 40, 73)
2044 1280100
(16, 16, 28, 23)(150, 10, 117, 19)
30332080.002299.78
(16, 11, 22)(150, 10, 78)
40532080.001799.83
(8, 16, 16, 32, 64)(10, 10, 10)
50431580.003399.67
(13, 12, 24, 32)(99, 109, 54)
60431280.005699.44
(8, 16, 32, 32)(150, 10, 150)
70451780.006799.33
(14, 16, 25, 23)(104, 150, 10, 21, 50)
80451580.012598.75
(16, 8, 32, 18)(150, 128, 10, 100, 10)
90151980.024997.51
(16)(105, 150, 108, 100, 47)
Table 4. The best, average, and worst accuracy values obtained by the PSO.
Table 4. The best, average, and worst accuracy values obtained by the PSO.
Images (Testing)
%
Best
%
Average
%
Worst
%
10-100-
2010099.6699.17
3099.7899.5999.22
4099.8399.5199.25
5099.6799.4999.20
6099.4499.1798.72
7099.3398.7298.00
8098.7598.0897.54
9097.5197.1596.14
Table 5. The best accuracy results and architectures obtained by the WOA. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
Table 5. The best accuracy results and architectures obtained by the WOA. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
%
Images for Testing
CLs
(Filters)
FCLs
(Neurons)
EpochBatch SizeErrorAccuracy (%)
10341980100
(9, 15, 21)(77, 84, 83, 27)
205520320100
(16, 16, 32, 32, 64)(150, 88, 150, 100, 50)
30422080.002299.78
(13, 16, 32, 27)(150, 143)
40542080.001799.83
(16, 13, 32, 32, 46)(150, 26, 136, 56)
50552080.003399.67
(16, 13, 32, 32, 64)(150, 80, 150, 100, 50)
605520160.005699.44
(16, 12, 32, 32, 64)(150, 137, 53, 55, 50)
70442080.006799.33
(16, 14, 30, 32)(150, 150, 114, 26)
80532080.012198.79
(16, 9, 32, 32, 54)(53, 150, 150)
90332080.022397.77
(14, 10, 23)(11, 96, 102)
Table 6. The best, average, and worst accuracy values obtained by the WOA.
Table 6. The best, average, and worst accuracy values obtained by the WOA.
Images (Testing)
%
Best
%
Average
%
Worst
%
1010099.9299.33
2010099.7699.50
3099.7899.5399.11
4099.8399.4699.17
5099.6799.4899.27
6099.4498.9498.27
7099.3398.7698.14
8098.7997.9497.24
9097.7797.1496.51
Table 7. The best accuracy results and architectures obtained by the BA. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
Table 7. The best accuracy results and architectures obtained by the BA. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
%
Images for Testing
CLs
(Filters)
FCLs
(Neurons)
EpochBatch SizeErrorAccuracy (%)
103314160100
(11, 10, 28)(121, 61, 63)
20331580.001799.83
(14, 15, 20)(66, 69, 34)
30442080.002299.78
(14, 13, 17, 31)(12, 43, 10, 75)
40432080.003399.67
(15, 8, 16, 16)(150, 150, 10)
50452080.002799.73
(16, 15, 32, 24)(150, 150, 150, 33, 28)
60452080.005099.50
(15, 16, 26, 32)(35, 150, 50, 36, 10)
70552080.007299.28
(8, 8, 32, 32, 64)(42, 150, 150, 100, 50)
80342080.010998.91
(16, 8, 32)(50, 29, 150, 100)
90241280.024597.55
(12, 11)(71, 75, 96, 25)
Table 8. The best, average, and worst accuracy values obtained by the BA.
Table 8. The best, average, and worst accuracy values obtained by the BA.
Images (Testing)
%
Best
%
Average
%
Worst
%
10-100-
2099.8399.7299.50
3099.7899.5499.22
4099.6799.4799.00
5099.7399.5399.33
6099.5099.2399.05
7099.2898.8998.33
8098.9198.1697.70
9097.5597.2396.84
Table 9. The best accuracy results and architectures obtained by the GWO. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
Table 9. The best accuracy results and architectures obtained by the GWO. CLs indicates the number of convolutional layers with their number of filters, and FCLs indicates the number of fully connected layers with their number of neurons.
%
Images for Testing
CLs
(Filters)
FCLs
(Neurons)
EpochBatch SizeErrorAccuracy (%)
104217320100
(13, 8, 27, 24)(122, 104)
20432080100
(10, 8, 23, 25)(42, 139, 32)
30321080.003399.67
(9, 8, 22)(38, 93)
40442080.002599.75
(16, 16, 16, 30)(150, 67, 106, 10)
50532080.003399.67
(8, 9, 32, 19, 64)(120, 81, 10)
60452080.006799.33
(9, 12, 16, 29)(63, 10, 53, 15, 15)
70431680.008199.19
(8, 8, 26, 32)(14, 102, 37)
80311580.017598.25
(16, 13)(48)
90141180.024197.59
(15)(107, 131, 117, 53)
Table 10. The best, average, and worst accuracy values obtained by the GWO.
Table 10. The best, average, and worst accuracy values obtained by the GWO.
Images (Testing)
%
Best
%
Average
%
Worst
%
1010099.9399.67
2010099.6299.00
3099.6799.4099.11
4099.7599.4799.17
5099.6799.3498.80
6099.3398.8498.22
7099.1998.7698.33
8098.2597.9197.62
9097.5997.1896.81
Table 11. Summary of accuracy results obtained by the metaheuristics.
Table 11. Summary of accuracy results obtained by the metaheuristics.
Images (Testing)
%
PSO
%
WOA
%
BA
%
GWO
%
1010099.9210099.93
2099.6699.7699.7299.62
3099.5999.5399.5499.40
4099.5199.4699.4799.47
5099.4999.4899.5399.34
6099.1798.9499.2398.84
7098.7298.7698.8998.76
8098.0897.9498.1697.91
9097.1597.1497.2397.18
Table 12. Summary of error results obtained by the metaheuristics.
Table 12. Summary of error results obtained by the metaheuristics.
%
Images (Testing)
PSOWOABAGWO
1000.000800.0007
200.00340.00240.00280.0038
300.00410.00470.00460.006
400.00490.00540.00530.0053
500.00510.00520.00470.0066
600.00830.01060.00770.0116
700.01280.01240.01110.0124
800.01920.02060.01840.0209
900.02850.02860.02770.0282
Table 13. Average results using Accuracy, Recall, Precision, and F1 Score.
Table 13. Average results using Accuracy, Recall, Precision, and F1 Score.
MetricPSOWOABAGWO
Accuracy 10099.9210099.93
Recall 97.0596.2899.7795.80
Precision 80.0782.6784.4781.38
F1 Score 86.1887.3190.5486.40
Table 14. Critical values for the Wilcoxon signed-rank test.
Table 14. Critical values for the Wilcoxon signed-rank test.
nα
0.020.050.10
9368
Table 15. Summary of Wilcoxon test results.
Table 15. Summary of Wilcoxon test results.
MethodsNegative Sum
(W−)
Positive Sum
(W+)
Test Statistic
(W)
Degrees of Freedom
(m)
W0 = Wα,m
BA
PSO
413398
BA
WOA
413398
BA
GWO
440098
PSO
WOA
34101098
PSO
GWO
395598
WOA
GWO
33101098
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Melin, P.; Sánchez, D.; Pulido, M.; Castillo, O. Comparative Study of Metaheuristic Optimization of Convolutional Neural Networks Applied to Face Mask Classification. Math. Comput. Appl. 2023, 28, 107. https://doi.org/10.3390/mca28060107

AMA Style

Melin P, Sánchez D, Pulido M, Castillo O. Comparative Study of Metaheuristic Optimization of Convolutional Neural Networks Applied to Face Mask Classification. Mathematical and Computational Applications. 2023; 28(6):107. https://doi.org/10.3390/mca28060107

Chicago/Turabian Style

Melin, Patricia, Daniela Sánchez, Martha Pulido, and Oscar Castillo. 2023. "Comparative Study of Metaheuristic Optimization of Convolutional Neural Networks Applied to Face Mask Classification" Mathematical and Computational Applications 28, no. 6: 107. https://doi.org/10.3390/mca28060107

Article Metrics

Back to TopTop