Next Article in Journal
Analysis of Available Solutions for the Improvement of Body Posture in Chairs
Next Article in Special Issue
Multirotor Unmanned Aerial Vehicle Configuration Optimization Approach for Development of Actuator Fault-Tolerant Structure
Previous Article in Journal
A Hybridized Flower Pollination Algorithm and Its Application on Microgrid Operations Planning
Previous Article in Special Issue
Hybrid Direction of Arrival Precoding for Multiple Unmanned Aerial Vehicles Aided Non-Orthogonal Multiple Access in 6G Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Swarm Intelligence with Deep Transfer Learning Driven Aerial Image Classification Model on UAV Networks

by
Saud S. Alotaibi
1,
Hanan Abdullah Mengash
2,
Noha Negm
3,4,
Radwa Marzouk
2,
Anwer Mustafa Hilal
5,*,
Mohamed A. Shamseldin
6,
Abdelwahed Motwakel
5,
Ishfaq Yaseen
5,
Mohammed Rizwanullah
5 and
Abu Sarwar Zamani
5
1
Department of Information Systems, College of Computing and Information System, Umm Al-Qura University, Mecca 24382, Saudi Arabia
2
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
3
Department of Computer Science, College of Science & Art at Mahayil, King Khalid University, Abha 62529, Saudi Arabia
4
Department of Mathematics and Computer Science, Faculty of Science, Menoufia University, Menoufia 32511, Egypt
5
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
6
Department of Mechanical Engineering, Faculty of Engineering and Technology, Future University in Egypt, New Cairo 11835, Egypt
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6488; https://doi.org/10.3390/app12136488
Submission received: 25 April 2022 / Revised: 20 June 2022 / Accepted: 22 June 2022 / Published: 26 June 2022
(This article belongs to the Special Issue Unmanned Aerial Vehicles)

Abstract

:
Nowadays, unmanned aerial vehicles (UAVs) have gradually attracted the attention of many academicians and researchers. The UAV has been found to be useful in variety of applications, such as disaster management, intelligent transportation system, wildlife monitoring, and surveillance. In UAV aerial images, learning effectual image representation was central to scene classifier method. The previous approach to the scene classification method depends on feature coding models with lower-level handcrafted features or unsupervised feature learning. The emergence of convolutional neural network (CNN) is developing image classification techniques more effectively. Due to the limited resource in UAVs, it can be difficult to fine-tune the hyperparameter and the trade-offs amongst computation complexity and classifier results. This article focuses on the design of swarm intelligence with deep transfer learning driven aerial image classification (SIDTLD-AIC) model on UAV networks. The presented SIDTLD-AIC model involves the proper identification and classification of images into distinct kinds. For accomplishing this, the presented SIDTLD-AIC model follows a feature extraction module using RetinaNet model in which the hyperparameter optimization process is performed by the use of salp swarm algorithm (SSA). In addition, a cascaded long short term memory (CLSTM) model is executed for classifying the aerial images. At last, seeker optimization algorithm (SOA) is applied as a hyperparameter optimizer of the CLSTM model and thereby results in enhanced classification accuracy. To assure the better performance of the SIDTLD-AIC model, a wide range of simulations are implemented and the outcomes are investigated in many aspects. The comparative study reported the better performance of the SIDTLD-AIC model over recent approaches.

1. Introduction

Unmanned aerial vehicles (UAV) are utilized as a cost-efficient and prompt methodology for taking remote sensing (RS) images. The boon of UAV technology involves least cost, small size, security, natural function, and, especially, the fast and on-demand acquisition of images [1]. The developments of UAV technologies have achieved the state that it can offer intense higher resolution RS images encircling lavish contextual and spatial information. This has allowed studies suggesting numerous original applications for UAV image examination, comprising disaster management, vegetation monitoring, object detection, detection and mapping of archaeological sites, oil and gas pipeline monitoring, and urban site analysis [2,3].
Aerial image classification methodologies grant distinct semantic categories that are usually established through exploiting changes in spatial deployments and structural forms for designing scenes [4]. In opposition with object or pixel related classifier methods, scene classification provides localization data from extensive aerial image which has apparent semantic data of the surfaces. Such methodologies are classified into three categories, which are: high level vision information, low level visual features, and mid-level visual representations [5,6]. Aerial scenes are differentiated by low level characteristics use, structural features, texture, spectral, and so on. Subsequently, low level feature vectors are pictorial ascriptions which can be derived globally or locally and are usually utilized for describing aerial scene images [7,8]. The typical low level feature methods are local binary patterns (LBP), Global Invariant Scale Transform (GIST), and color histogram Scale Invariant Feature Transform (SIFT). Mid-level analytical methods try to advance complete scene illustrations through conveying high order statistical outlines which are created by deriving local visual qualities [9]. The common processing pipeline derives local image patches, and they are programmed as local signals; therefore, creating a complete mid-level depiction of the aerial scenes. The familiar mid-level procedure is Bag of Visual Words (BoVW).
Deep learning (DL) procedures like Convolutional Neural Networks (CNNs) were broadly recognized as a notable approach for numerous computer vision applications (classification, image or video recognition, and detection), and have revealed amazing outcomes in various applications [10]. Therefore, there comes numerous advantages to stopping from utilizing DL methods in emergency response and calamity management applications to restore crucial data in a timely manner and permitting superior research and response in the course of time-critical circumstances, and supporting the decision-making processes [11]. Although CNNs were rising successfully at several classification roles via transfer learning (TL), their interpretation speed on implanted platforms, like those discovered on-board UAVs, is hampered by the high computational cost, which may acquire and the model size of these networks is prohibitive from a memory standpoint for these entrenched gadgets [12]. At the same time, most of the earlier works do not consider hyperparameter tuning process into account.
This article focuses on the design of swarm intelligence with deep transfer learning driven aerial image classification (SIDTLD-AIC) model on UAV networks. The presented SIDTLD-AIC model follows a feature extraction module using RetinaNet model, in which the hyperparameter optimization process is performed by the use of salp swarm algorithm (SSA). The SSA is chosen as it avoids the local optimal constraints, thus achieving a smooth balance between exploration and exploitation. In addition, a cascaded long short term memory (CLSTM) model is executed for classifying the aerial images. At last, seeker optimization algorithm (SOA) is applied as a hyperparameter optimizer of the CLSTM model and thereby results in enhanced classification accuracy. To assure the better performance of the SIDTLD-AIC model, a wide range of simulations are executed and the outcomes are investigated in various aspects.

2. Related Works

Haq et al. [13] applied DL based supervised image classification model and images gathered using UAV for the forest region classification. The DL technique based stacked Autoencoder (SAE) has shown remarkable potential with respect to the assessment of forest areas and image classification. The experiment result shows that DL technique provides improved performance than other machine learning approaches. The researchers in [14] address the shortcoming of multi-labeling UAV images, usually considered by a higher level of dataset content, by presenting a novel technology based on CNN. They are employed as a means to produce an accurate representation of the query images that are analyzed afterward sub-dividing them into a grid of tiles. The multi-label classification process is implemented by combining a radial basis function neural network and a multi-labeling layer comprised of threshold operation. The researchers in [15] proposed a DL algorithm for classifying UAV images derived from the location and sensor of earth’s surface. Initially, the labelled and unlabelled UAV images are fed to a pre-trained CNN to generate deep feature representation. Next, we learned strong domain-invariant features with a further network comprised of two fully connected layers.
Rajagopal et al. [16] developed a new optimum DL-based scene classification algorithm captured by UAV. The suggested method includes a residual network-based features extraction (RNBFE) that extract feature from the convolutional layer of a DRN system. Furthermore, the various parameters result in configuration errors because of parameter tuning. Hence, self-adoptive global best harmony search (SGHS) approach is applied to tune the parameter of the presented model. The researchers in [17] present a multi-objective optimization algorithm to evolve deep CNN for scene classification that generates the non-dominant solution in an automatic manner at the Pareto front. Then, we used two sets of benchmark data sets for testing the effectiveness of the scene classification algorithm and making an extensive analysis. Pustokhina et al. [18,19] presented an energy-effective cluster-based UAV system using DL based scene classification model. The suggested method includes a clustering with parameter tuned residual network (C-PTRN) system that operates on two primary processes scene classification and cluster construction.

3. The Proposed Model

In this article, an automated SIDTLD-AIC method was established for the proper identification and classification of images into distinct kinds on UAV networks. The presented SIDTLD-AIC model follows a feature extraction module using RetinaNet model in which the hyperparameter optimization process is performed by the use of SSA. Next, the SOA-CLSTM model is applied to classify the aerial images. Figure 1 depicts the block diagram of SIDTLD-AIC technique.

3.1. Feature Extraction Using RetinaNet Model

Transfer learning model is applied to enhance the efficiency of the DL model by the use of labeled data. It learns and employs many source processes for enhancing the learning process in relevant domains. It encompasses pre-training approaches which is trained on large scale dataset and is retrained at varying levels of the model on a small training set. The preliminary layer of the pre-training network can be modified upon requirement. The final layer of the model’s hyperparameters can be tuned for learning the abilities on new datasets. In this work, the RetinaNet based TL model is applied for deriving feature vectors. An input map was inspired by the individual layer still accomplishing the resulting map. The CNN is designed in an order of layers. Consider R h × w × c ( h : height, w : width, c : channel) are RGB images. Each layer gets X and the set of variables W as input as well as output images y R h × w × c , for example, y = f X ,   W . This makes an activation map to demonstrate the reaction of that filter at each spatial region. For calculating the input X with set of filters W R h ¯ x w ¯ × c ¯ × c and add a bias b R c as follows.
y i j k = f b k + i = 1 h ¯ j = 1 w ¯ d = 1 c W i j d k × X i + i , j + j , d .
Next, the max-pooling layer is employed to decrease the computation and parameter with the decreased size of imputing shapes. It evaluates the maximum response of every image channel from h ¯ × W sub-windows that implement as sub-sampling function. It can be expressed in the following:
y i j k = max , 1 < i < h ˜ 1 < j < w ˜ X i + i j + j , k .
Finally, fully connected (FC) layer is a set of layers that integrate the data extracted by previous layer (feature). This layer gets an input X , processes them, and the last FC layer generates one dimensional vector of size. RetinaNet mainly consists of [20] two fully convolution network (FCNs), ResNet, and feature pyramid network (FPN). The ResNet employs network layer. The widely employed types of network layers are 50_, 101_, and 152_layers. The 101_layer with optimal trained efficacy is chosen. It could eliminate the structure of echocardiography with ResNet and, after, keep them to following sub-network. An FPN is an approach for efficiently eliminating the feature of each dimension from image with a conventional CNN architecture.
Focal loss: it can be improved version of the binary cross entropy (CE), the loss expression is given below:
C E   p ,   y = log p i f   y = 1 log l p o t h e r w i s e ,
In Equation (3), y 1 , + 1 indicates the ground truth type and p 0 ,   1 represents the prediction probability to type y = 1 .
p t = p , i f   y = l l p , o t h e r w i s e
The preceding formula can be abbreviated as follows:
C E   p ,   y = C E p t =   log   p t
In order to resolve the problem of data imbalance among the negative and positive samples, the new process is changed into the succeeding process:
C E p t = α t   log   p t .
Among them,
α t = α , i f   y = 1 , l α o t h e r w i s e ,
Here, α 0 ,   1 represents the weight factor. In order to resolve the problem, the concentrating variable C was determined to obtain the final process of focal loss:
F L p t = α t ( 1 p t ) γ   log   p t .

3.2. Hyperparameter Optimization: SSA

In this work, the hyperparameters of the RetinaNet model such as number of epochs, batch size, learning rate, and momentum are adjusted by the design of SSA. The SSA is simulated from the aggregation performance of salps that procedure a chain of salps and then hunt and move. The salp chain was developed from two kinds of salps, leader and follower [21]. The leader is the salp at the head of chains. Individual salps at the back of chains are followers. During the salp technique, food source F was determined as the individual with optimum fitness amongst every individual. The food source of tth order is F t . The steps of SSA technique are provided under.
1.
Initialization of the population. In order for every individual, the places were arbitrary numbers amongst the upper as well as lower limits. They then compute fitness of every individual and sorted them. An individual with minimal fitness is the food source F t . t = 1 , since one iteration was ended.
2.
The population place was upgraded. The leader place was upgraded as:
x j i t + 1 = F j t + c 1 u b j l b j c 2 + l b j c 3 0.5 F j t c 1 u b j l b j c 2 + l b j c 3 < 0.5
where i = 1 , i.e., the count of leaders is 1. It ranks primary from the populations. j = 1 , 2 D . F j , u b j , and l b j are F t , u b , and l b from the jth dimensional correspondingly. c 2 and c 3 implies the arbitrary numbers from zero and one. c 2 affects the step length of leader movement. c 3 defines if the leader moves forward/backward to food sources. T signifies the maximal number of iterations. c 1 refers the co-efficient of moving length.
c 1 = 2 e ( 4 t / T ) 2
The place of follower is:
x j i t + 1 = 0.5 x j i 1 t + x j i   t
where i 2 , and is the sequence of followers from the population. j = 1 , 2 D .
3.
Compute the fitness of every upgraded individual. The sort of individuals. Upgrade F t . Improve t by 1.
4.
If the iteration accuracy condition was attained or t = T , the iteration terminates; or else, go to ( 2 ) to remain the iteration.

3.3. Image Classification Using Optimal CLSTM Model

In the final stage, the optimal CLSTM model is utilized to recognize different types of classes that exist in the aerial images [22]. A recurrent neural network (RNN) is a kind of DL method which is depending on existing input and the preceding input. In general, it is suitable for the scenario where the dataset has a consecutive correlation. But while handling a longer series of datasets, there exists an exploiting and vanishing gradient problem. In order to resolve this problem, a long short term memory (LSTM) is utilized that has an internal memory state which adds forget gate. The gate controls the time dependency and the effects of preceding inputs. Bidirectional long short term memory (BiLSTM) and bidirectional RNN (BiRNN) are other variations that reflect preceding input and assume the upcoming input of a certain time frame. This work, can present the BiLSTM RNN and cascaded uni-directional LSTM models. The method comprises the initial layer of bi-directional RNN integrated with uni-directional RNN layer. The bi-directional LSTM comprises forward and backward tracks to learn patterns in two directions.
O n f 1 ,   h n f 1 ,   i n f 1 = L f 1 i n 1 f 1 ,   h n 1 f 1 ,   x n   : P f 1 ,
O n b 1   ,   h n b 1   ,   i n b 1 = L b 1   i n 1 b 1 ,   h n 1 b 1 ,   x n : P b 1 ,
Equations (12) and (13) show the operation of forwarding and backward tracks. Figure 2 depicts the framework of LSTM.
From the equation, O n f 1 , h n f 1 , i n f 1 and O n b 1 , h n b 1 , i n b 1 indicate the output, the hidden state, and the internal state of the existing state for forwarding and backward LSTM tracks correspondingly. x n denotes the sequential input, P indicates the LSTM cell variable. The output from these two tracks is integrated as in Equation (14) and forwarded into the next layers.
O n 1 = O n f 1 + O N n + 1 b 1 .
Bi-directional RNN and uni-directional RNN transform information into an abstract form and assist in learning spatial dependency. The output from the uni-directional layer can be attained by the following equation.
O n l , h n l , i n l = LSTM l   i n 1 l ,   h n 1 l ,   O n l 1 ; P l ,  
Now, the output from the lower layer O n l 1 is integrated with preceding internal state i n 1 l and hidden state h n 1 l for obtaining output O n l of layer l , and P l indicates a variable of the LSTM cell. The input dataset comprises a series of instances x 1 ,   x 2 , ,   x N , while every feature x n is regarded at time n   n = 1 , 2 ,   ,   N . The information is mainly classified into windows of time segment N and fed into the cascading LSTM. We attain predicted score vectors for every time step O 1   L ,   O 2 L , ,   O N L at the output. The entire prediction score can be attained by integrating the predictive score vector for the window N . The combination of scores can be implemented by using the sum rule as demonstrated in Equation (16) that implements well than other methodologies. Finally, the predictive score is transformed into probability using a softmax layer over Y .
Y = 1 N n = 1 N O n L .
We cascade LSTM to simulate incremental change of n time steps, and every LSTM is utilized for estimating the increment for one time step. In this work, θ-increment learning method learns increment of parameters using the cascaded LSTM network to gain higher frequency approximation, and θ represents the targeted parameter to be calculated.
In order to optimally elect the hyperparameter values of the CLSTM model, the SOA is exploited. In SOA, all the seekers have a central location vector c , viz., the initial position for finding upcoming solutions, and it is regarded as estimated value E x . Furthermore, all the seekers have a searching radius r regarded as the E n , a trust level μ as membership degree, and a searching direction d . Next, the seeker with some level of trust followed a potential direction and randomly moves towards the second point (novel candidate solution) in some searching radius from their existing location. In every time step t , the search decision-making is carried out for evaluating the four variables, and the seeker moves toward the novel location x t + 1 . The updating location from the central location can be defined as y -conditional cloud generator [23]:
x i j t + 1 = c i j t + d i j t × r i j t × ln μ i
Here “ i ” refers to the subscript index of seeker, and “ j ” indicates the subscript index of parameter dimension. The pseudo-code of the SOA is given in Algorithm 1.
Algorithm 1: Pseudocode of SOA
t 0
Initialized generation of S position
{ x i t | x i t = x i 1 , x i 2 , , x i D ,   i = 1 , , S ,   t = 0 } Uniformly and randomly in the parameters.
Estimate all the seekers: Compute the fitness.
Searching techniques provide search variables involving central location vector, searching direction, searching radius, and trust degree.
Update new location of all the seekers is evaluated.
t t + 1
When t < T max , then Go to 3; otherwise, End.
Instinctively, central location vector c is fixed to existing location x t . Similar to particle swarm optimization (PSO), all the seekers contain a memory stored in its optimal location p and a global optimal location g accomplished by communicating with neighboring seekers. Every seeker is categorized into k class in the subscript index, and the seeker in a similar class belongs to virtual neighbors. Therefore, g is established in the virtual neighbors.
c = x t + r 1 1 p t x t + r 2 2 g t x t
Now r 1 ,   r 2 indicates the cognitive and social learning rates, correspondingly. 1 and 2 denotes the real number randomly and uniformly selected within the range of [0, 1]. In every experiment carried out in the study, r 1 = 1 , r 2 = 1 , and k = 3 .
Generally, all the seekers have four significant directions, named local spacial direction d l s , local temporal direction d l t , global spacial direction d g s , and global temporal direction d g t , correspondingly.
d l t = s i g n   x t x t 1   i f   f i t x t f i t x t 1 s i g n   x t 1 x t   i f   f i t x t < f i t x t 1
d l s = s i g n x t x t
d g t = s i g n p t x t
d g s = s i g n g t x t
From the above equation, sign indicates signum function, x t represent the location of the seekers with the maximum fitness in a neighbor region, fit x t denotes the fitness function (FF) of x t . Next, searching direction is allocated based on the four directions.
d = s i g n ( ω s i g n f i t x t f i t x t 1 x t x t 1 + r 1 ϕ 1 p t x t + r 2 ϕ 2 g t x t )
In Equation (7), ω indicates the inertia weight that is fixed to ω = T m a x t / T max . Now, ϕ 1 and ϕ 2 indicates real numbers randomly and uniformly selected within [0, 1].
Search Radius is essential, but challenging, to reasonably provide searching radius. For unimodal optimization problems, the performance is comparatively oblivious to searching radius to some extent. However, for multi-modal problems, various searching radii might lead to various performances of model particularly while handling variety of problems.
The μ variable is considered a quality assessment of location. It is equivalent to the fitness of x t or the index of ascensive sorting order of the fitness of x t . Especially, the global optimal location has the maximal μ max = 1.0 , when another location has a μ < 1.0 .
μ = μ   max   S S n S 1 μ   max   μ   min
Here, S n indicates the sequential value of x t afterward arranging the finesses of neighboring seekers in ascending sequence, μ   max   and μ   min indicates the maximal and the minimal μ . We adapted μ   max   = 1.0 , and μ   min   = 0.2 .
The SOA method develops a fitness function (FF) to accomplish better classification accuracy. It describes a positive integer to characterize the improved performance of the candidate solution. In this work, the reduction of the classification error rate is regarded as the FF, as shown in Equation (25).
f i t n e s s x i = C l a s s i f i e r   E r r o r   R a t e x i = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s 100

4. Experimental Validation

The performance validation of the proposed model is carried out using the UCM dataset [24]. The dataset contains a total of 2100 images and 21 classes (agricultural, airplane, baseballdiamond, beach, buildings, chaparral, denseresidential, forest, freeway, golfcourse, harbor, intersection, mediumresidential, mobilehomepark, overpass, parkinglot, river, runway, sparseresidential, storagetanks, and tenniscourt). It includes a total of 100 images under each class. The images were manually extracted from large images from the USGS National Map Urban Area Imagery collection for various urban areas around the country. The pixel resolution of this public domain imagery is 1 foot. Each image measures 256 × 256 pixels. For experimental validation, the dataset is split into 70% of training set and 30% of testing set, i.e., 70 images from each class for training and remaining 30 images for testing purposes. Figure 3 showcases the sample images of UCM dataset.
Figure 4 illustrates the confusion matrices provided by the SIDTLD-AIC model on 70% of UCM datasets as training datasets. The results indicated that the SIDTLD-AIC model has effectually categorized all the 21 classes.
Table 1 reports the overall classification outcomes of the SIDTLD-AIC model 70% of UCM datasets as training datasets. The results inferred that the SIDTLD-AIC model has accomplished enhanced classifier outcomes on all class labels. For instance, the SIDTLD-AIC model has recognized class 1 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.39%, 90.41%, 97.06%, 93.62%, and 98.27%, respectively. Along with that, the SIDTLD-AIC method has recognized class 3 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.66%, 98.53%, 94.37%, 96.40%, and 97.11%, correspondingly. Moreover, the SIDTLD-AIC system has recognized class 13 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.32%, 91.67%, 94.29%, 92.96%, and 96.89%, correspondingly. Furthermore, the SIDTLD-AIC approach has recognized class 16 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.66%, 97.10%, 95.71%, 96.40%, and 97.76%, respectively. Lastly, the SIDTLD-AIC method has recognized class 20 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.73%, 95.71%, 98.53%, 97.10%, and 99.16%, correspondingly.
Figure 5 showcases the confusion matrices provided by the SIDTLD-AIC approach on 30% of UCM datasets as testing datasets. The results point out that the SIDTLD-AIC methodology has effectually categorized all the 21 classes.
Table 2 demonstrates the overall classification outcomes of the SIDTLD-AIC method on 30% of UCM datasets as testing datasets. The results exposed that the SIDTLD-AIC model has accomplished higher classifier outcomes on all class labels. For instance, the SIDTLD-AIC method has recognized class 1 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.68%, 100%, 93.75%, 96.77%, and 96.82%, respectively. Next, the SIDTLD-AIC model has recognized class 3 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.52%, 96.43%, 93.10%, 94.74%, and 96.41%, correspondingly. Furthermore, the SIDTLD-AIC system has recognized class 13 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.21%, 93.10%, 90%, 91.53%, and 94.71%, respectively. Moreover, the SIDTLD-AIC methodology has recognized class 16 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.52%, 100%, 90%, 94.74%, and 94.87%, respectively. Finally, the SIDTLD-AIC model has recognized class 20 samples with a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.21%, 88.57%, 96.88%, 92.54%, and 98.10%, correspondingly.
The training accuracy (TA) and validation accuracy (VA) attained by the SIDTLD-AIC model on UCM dataset is demonstrated in Figure 6. The experimental outcome implied that the SIDTLD-AIC model has gained maximum values of TA and VA. In specific, the VA seemed higher than TA.
The training loss (TL) and validation loss (VL) achieved by the SIDTLD-AIC model on UCM dataset are established in Figure 7. The experimental outcome inferred that the SIDTLD-AIC model has been able least values of TL and VL. In specific, the VL seemed that lower than TL.
A brief precision-recall examination of the SIDTLD-AIC method on UCM dataset is portrayed in Figure 8. By observing the figure, it can be noticed that the SIDTLD-AIC method has been able maximal precision-recall performance under all classes.
A detailed ROC investigation of the SIDTLD-AIC approach to UCM dataset is represented in Figure 9. The results indicated that the SIDTLD-AIC model has exhibited its ability in categorizing different classes on the UCM dataset.
Figure 10 depicts the average image classification results of the SIDTLD-AIC model on 70% of UCM datasets as training datasets and 30% of UCM datasets as testing datasets. The figure shows that the SIDTLD-AIC model has resulted in better classification results under both aspects. On applied 70% of UCM datasets as training datasets, the SIDTLD-AIC model has resulted in average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 96.60%, 95.89%, 95.86%, 95.85%, and 97.80%, respectively. Likewise, on applied 30% of UCM datasets as testing datasets, the SIDTLD-AIC model has resulted in average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a n of 99.50%, 94.93%, 94.67%, 94.70%, and 97.15%, respectively.
Figure 11 illustrates a comparative a c c u y examination of the SIDTLD-AIC model with recent models. The experimental values implied that the DL-PlacesNet and DL-VGG-VD19 models have shown lower values of a c c u y . Moreover, the DL-VGG-VD16, DL-VGG-M, DL-VGG-F, DL-CaffeNet, and DL-AlexNet models have resulted to closer a c c u y values. Then, the DL-VGG-S and DL based multiobjective PSO (DL-MOPSO) techniques have reached reasonable a c c u y values of 95.24% and 95.81%. Though the DL-C-PTRN model has resulted in considerable a c c u y of 98.96%, the SIDTLD and SIDTLD+SSA models have accomplished near optimal a c c u y of 98.98% and 99.01%. However, the SIDTLD-AIC model has accomplished superior outcome with maximum a c c u y of 99.50%.
Finally, a computation time (CT) assessment of the SIDTLD-AIC model with recent models is carried out in Figure 12. The experimental values implied that the DL-PlacesNet, DL-VGG-VD-19, DL-VGG-VD-16, and DL-VGG-S approaches have obtained increased CT values. Followed by, the DL-VGG-F, DL-CaffeNet, and DL-AlexNet models have reached moderately reduced CT values. The DL-MOPSO and DL-C-PTRN models have accomplished reasonable CT of 135s and 95s, respectively. Meanwhile, the SIDTLD and SIDTLD+SSA models have attained CT of 67s and 54s, respectively. Finally, the SIDTLD-AIC model has outperformed other methods with minimal CT of 40s. The results implied that the SIDTLD-AIC model has gained enhanced classification performance due to the inclusion of SSA and SOA based hyperparameter optimizers. From the above results and discussion, it can be stated that the SIDTLD-AIC model has accomplished enhanced image classification results on the UAV networks.

5. Conclusions

In this article, an automated SIDTLD-AIC technique was established for the proper identification and classification of images into distinct kinds on UAV networks. The presented SIDTLD-AIC model follows a feature extraction module using RetinaNet model in which the hyperparameter optimization process is performed by the use of SSA. Next, the SOA-CLSTM model is applied to classify the aerial images. For assuring the better performance of the SIDTLD-AIC method, a wide range of simulations are executed and the outcomes are investigated in various aspects. The comparative study reported the better performance of the SIDTLD-AIC model over recent approaches with maximum accuracy of 99.50%. Thus, the presented SIDTLD-AIC model can be exploited for aerial image classification in real time environment such as vegetation mapping, crop classification, disaster management, weather prediction, etc. In future, hybrid metaheuristics should be utilized for improving the overall classification performance. Furthermore, the proposed model can be extended to real-time large-scale databases in future. Moreover, the investigation of the performance using statistical analysis can be done in our future work.

Author Contributions

Conceptualization, S.S.A. and N.N.; methodology, H.A.M.; software, R.M.; validation, A.M.H., R.M. and M.R.; investigation, A.M.H., M.A.S.; resources, M.R.; data curation, R.M.; writing—original draft preparation, S.S.A., H.A.M., N.N., I.Y.; writing—review and editing, A.M., M.A.S., A.S.Z.; visualization, A.S.Z.; supervision, S.S.A.; project administration, R.M.; funding acquisition, H.A.M. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups Project under grant number (42/43). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R114), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: (22UQU4210118DSR21).

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through contributions of all authors. All authors have given approval to the final version of the manuscript.

References

  1. Choi, S.K.; Lee, S.K.; Kang, Y.B.; Seong, S.K.; Choi, D.Y.; Kim, G.H. Applicability of image classification using deep learning in small area: Case of agricultural lands using UAV image. J. Korean Soc. Surv. Geod. Photogramm. Cartogr. 2020, 38, 23–33. [Google Scholar]
  2. Tetila, E.C.; Machado, B.B.; Astolfi, G.; de Souza Belete, N.A.; Amorim, W.P.; Roel, A.R.; Pistori, H. Detection and classification of soybean pests using deep learning with UAV images. Comput. Electron. Agric. 2020, 179, 105836. [Google Scholar] [CrossRef]
  3. Öztürk, A.E.; Erçelebi, E. Real UAV-bird image classification using CNN with a synthetic dataset. Appl. Sci. 2021, 11, 3863. [Google Scholar] [CrossRef]
  4. Ammour, N.; Alhichri, H.; Bazi, Y.; Benjdira, B.; Alajlan, N.; Zuair, M. Deep learning approach for car detection in UAV imagery. Remote Sens. 2017, 9, 312. [Google Scholar] [CrossRef] [Green Version]
  5. Bashmal, L.; Bazi, Y.; Al Rahhal, M.M.; Alhichri, H.; Al Ajlan, N. UAV image multi-labeling with data-efficient transformers. Appl. Sci. 2021, 11, 3974. [Google Scholar] [CrossRef]
  6. Anwer, M.H.; Hadeel, A.; Fahd, N.A.-W.; Mohamed, K.N.; Abdelwahed, M.; Anil, K.; Ishfaq, Y.; Abu Sarwar, Z. Fuzzy cognitive maps with bird swarm intelligence optimization-based remote sensing image classification. Comput. Intell. Neurosci. 2022, 2022, 4063354. [Google Scholar]
  7. Abunadi, I.; Althobaiti, M.M.; Al-Wesabi, F.N.; Hilal, A.M.; Medani, M.; Hamza, M.A.; Rizwanullah, M.; Zamani, A.S. Ederated learning with blockchain assisted image classification for clustered UAV networks. Comput. Mater. Contin. 2022, 72, 1195–1212. [Google Scholar]
  8. Li, J.; Yan, D.; Luan, K.; Li, Z.; Liang, H. Deep learning-based bird’s nest detection on transmission lines using UAV imagery. Appl. Sci. 2020, 10, 6147. [Google Scholar] [CrossRef]
  9. Youme, O.; Bayet, T.; Dembele, J.M.; Cambier, C. Deep Learning and Remote Sensing: Detection of Dumping Waste Using UAV. Procedia Comput. Sci. 2021, 185, 361–369. [Google Scholar] [CrossRef]
  10. Mittal, P.; Singh, R.; Sharma, A. Deep learning-based object detection in low-altitude UAV datasets: A survey. Image Vis. Comput. 2020, 104, 104046. [Google Scholar] [CrossRef]
  11. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. Vehicle detection from UAV imagery with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–21, in press. [Google Scholar] [CrossRef] [PubMed]
  12. Huang, H.; Deng, J.; Lan, Y.; Yang, A.; Deng, X.; Zhang, L.; Wen, S.; Jiang, Y.; Suo, G.; Chen, P. A two-stage classification approach for the detection of spider mite-infested cotton using UAV multispectral imagery. Remote Sens. Lett. 2018, 9, 933–941. [Google Scholar] [CrossRef]
  13. Haq, M.A.; Rahaman, G.; Baral, P.; Ghosh, A. Deep learning based supervised image classification using UAV images for forest areas classification. J. Indian Soc. Remote Sens. 2021, 49, 601–606. [Google Scholar] [CrossRef]
  14. Zeggada, A.; Melgani, F.; Bazi, Y. A deep learning approach to UAV image multilabeling. IEEE Geosci. Remote Sens. Lett. 2017, 14, 694–698. [Google Scholar] [CrossRef]
  15. Bashmal, L.; Bazi, Y. Learning robust deep features for efficient classification of UAV imagery. In Proceedings of the 2018 1st International Conference on Computer Applications & Information Security (ICCAIS), Riyadh, Saudi Arabia, 4–6 April 2018; pp. 1–4. [Google Scholar]
  16. Rajagopal, A.; Ramachandran, A.; Shankar, K.; Khari, M.; Jha, S.; Lee, Y.; Joshi, G.P. Fine-tuned residual network-based features with latent variable support vector machine-based optimal scene classification model for unmanned aerial vehicles. IEEE Access 2020, 8, 118396–118404. [Google Scholar] [CrossRef]
  17. Rajagopal, A.; Joshi, G.P.; Ramachandran, A.; Subhalakshmi, R.T.; Khari, M.; Jha, S.; Shankar, K.; You, J. A deep learning model based on multi-objective particle swarm optimization for scene classification in unmanned aerial vehicles. IEEE Access 2020, 8, 135383–135393. [Google Scholar] [CrossRef]
  18. Pustokhina, I.V.; Pustokhin, D.A.; Kumar Pareek, P.; Gupta, D.; Khanna, A.; Shankar, K. Energy-efficient cluster-based unmanned aerial vehicle networks with deep learning-based scene classification model. Int. J. Commun. Syst. 2021, 34, e4786. [Google Scholar] [CrossRef]
  19. Outay, F.; Mengash, H.A.; Adnan, M. Applications of unmanned aerial vehicle (UAV) in road safety, traffic and highway infrastructure management: Recent advances and challenges. Transp. Res. Part A Policy Pract. 2020, 141, 116–129. [Google Scholar] [CrossRef]
  20. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. Automatic ship detection based on RetinaNet using multi-resolution Gaofen-3 imagery. Remote Sens. 2019, 11, 531. [Google Scholar] [CrossRef] [Green Version]
  21. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  22. Yadav, R.K.; Bhattarai, B.; Jiao, L.; Goodwin, M.; Granmo, O.C. Indoor Space Classification Using Cascaded LSTM. In Proceedings of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November 2020; pp. 1110–1114. [Google Scholar]
  23. Shafik, M.B.; Chen, H.; Rashed, G.I.; El-Sehiemy, R.A. Adaptive multi objective parallel seeker optimization algorithm for incorporating TCSC devices into optimal power flow framework. IEEE Access 2019, 7, 36934–36947. [Google Scholar] [CrossRef]
  24. Yang, Y.; Newsam, S. Bag-Of-Visual-Words and Spatial Extensions for Land-Use Classification. In Proceedings of the ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems (ACM GIS), San Jose, CA, USA, 2–5 November 2010. [Google Scholar]
Figure 1. Block diagram of SIDTLD-AIC technique.
Figure 1. Block diagram of SIDTLD-AIC technique.
Applsci 12 06488 g001
Figure 2. Infrastructure of LSTM.
Figure 2. Infrastructure of LSTM.
Applsci 12 06488 g002
Figure 3. Samples-UCM Dataset.
Figure 3. Samples-UCM Dataset.
Applsci 12 06488 g003
Figure 4. Confusion matrix of SIDTLD-AIC technique on 70% of UCM datasets as training datasets.
Figure 4. Confusion matrix of SIDTLD-AIC technique on 70% of UCM datasets as training datasets.
Applsci 12 06488 g004
Figure 5. Confusion matrix of SIDTLD-AIC technique on 30% of UCM datasets as testing datasets.
Figure 5. Confusion matrix of SIDTLD-AIC technique on 30% of UCM datasets as testing datasets.
Applsci 12 06488 g005
Figure 6. TA and VA analysis of SIDTLD-AIC technique on UCM dataset.
Figure 6. TA and VA analysis of SIDTLD-AIC technique on UCM dataset.
Applsci 12 06488 g006
Figure 7. TL and VL analysis of SIDTLD-AIC technique on UCM dataset.
Figure 7. TL and VL analysis of SIDTLD-AIC technique on UCM dataset.
Applsci 12 06488 g007
Figure 8. Precision-recall curve analysis of SIDTLD-AIC technique on UCM dataset.
Figure 8. Precision-recall curve analysis of SIDTLD-AIC technique on UCM dataset.
Applsci 12 06488 g008
Figure 9. ROC curve analysis of SIDTLD-AIC technique on UCM dataset.
Figure 9. ROC curve analysis of SIDTLD-AIC technique on UCM dataset.
Applsci 12 06488 g009
Figure 10. Average analysis of SIDTLD-AIC technique with various measures.
Figure 10. Average analysis of SIDTLD-AIC technique with various measures.
Applsci 12 06488 g010
Figure 11. Accuracy analysis of SIDTLD-AIC technique with existing methods.
Figure 11. Accuracy analysis of SIDTLD-AIC technique with existing methods.
Applsci 12 06488 g011
Figure 12. CT analysis of SIDTLD-AIC technique with existing approaches.
Figure 12. CT analysis of SIDTLD-AIC technique with existing approaches.
Applsci 12 06488 g012
Table 1. Result analysis of SIDTLD-AIC technique with various measures on 70% of UCM datasets as training datasets.
Table 1. Result analysis of SIDTLD-AIC technique with various measures on 70% of UCM datasets as training datasets.
Training Phase (70%)
Class LabelsAccuracyPrecisionRecallF-ScoreGeometric Mean
099.7398.4895.5997.0197.73
199.3990.4197.0693.6298.27
299.5296.8892.5494.6696.13
399.6698.5394.3796.4097.11
499.86100.0097.1098.5398.54
599.3995.6591.6793.6295.64
699.5997.2294.5995.8997.19
799.7397.1897.1897.1898.51
899.3993.5194.7494.1297.16
999.7397.1097.1097.1098.47
1099.5997.0694.2995.6597.03
1199.6695.5296.9796.2498.37
1299.7398.6396.0097.3097.94
1399.3291.6794.2992.9696.89
1499.4693.9493.9493.9496.78
1599.5994.2097.0195.5998.36
1699.6697.1095.7196.4097.76
1799.5994.3797.1095.7198.40
1899.8096.00100.0097.9699.89
1999.5994.5997.2295.8998.46
2099.7395.7198.5397.1099.16
Average99.6095.8995.8695.8597.80
Table 2. Result analysis of SIDTLD-AIC technique with various measures on 30% of UCM datasets as testing datasets.
Table 2. Result analysis of SIDTLD-AIC technique with various measures on 30% of UCM datasets as testing datasets.
Testing (30%)
Class LabelsAccuracyPrecisionRecallF-ScoreGeometric Mean
0100.00100.00100.00100.00100.00
199.68100.0093.7596.7796.82
299.5291.67100.0095.6599.75
399.5296.4393.1094.7496.41
499.68100.0093.5596.6796.72
599.3792.8692.8692.8696.20
699.8496.30100.0098.1199.92
799.3796.3089.6692.8694.61
899.5295.6591.6793.6295.66
999.2186.11100.0092.5499.58
1099.6896.6796.6796.6798.24
1199.84100.0097.0698.5198.52
1299.0585.1992.0088.4695.60
1399.2193.1090.0091.5394.71
1499.2196.7788.2492.3193.85
1599.3791.4396.9794.1298.23
1699.52100.0090.0094.7494.87
1799.8496.88100.0098.4199.92
1899.68100.0092.8696.3096.36
1999.2189.6692.8691.2396.12
2099.2188.5796.8892.5498.10
Average99.5094.9394.6794.7097.15
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

S. Alotaibi, S.; Abdullah Mengash, H.; Negm, N.; Marzouk, R.; Hilal, A.M.; Shamseldin, M.A.; Motwakel, A.; Yaseen, I.; Rizwanullah, M.; Zamani, A.S. Swarm Intelligence with Deep Transfer Learning Driven Aerial Image Classification Model on UAV Networks. Appl. Sci. 2022, 12, 6488. https://doi.org/10.3390/app12136488

AMA Style

S. Alotaibi S, Abdullah Mengash H, Negm N, Marzouk R, Hilal AM, Shamseldin MA, Motwakel A, Yaseen I, Rizwanullah M, Zamani AS. Swarm Intelligence with Deep Transfer Learning Driven Aerial Image Classification Model on UAV Networks. Applied Sciences. 2022; 12(13):6488. https://doi.org/10.3390/app12136488

Chicago/Turabian Style

S. Alotaibi, Saud, Hanan Abdullah Mengash, Noha Negm, Radwa Marzouk, Anwer Mustafa Hilal, Mohamed A. Shamseldin, Abdelwahed Motwakel, Ishfaq Yaseen, Mohammed Rizwanullah, and Abu Sarwar Zamani. 2022. "Swarm Intelligence with Deep Transfer Learning Driven Aerial Image Classification Model on UAV Networks" Applied Sciences 12, no. 13: 6488. https://doi.org/10.3390/app12136488

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop