Next Article in Journal
A Dynamically Reconfigurable Autonomous Underwater Robot for Karst Exploration: Design and Experiment
Next Article in Special Issue
Using Deep Neural Networks for Human Fall Detection Based on Pose Estimation
Previous Article in Journal
Three-Dimensional-Printed Mechanical Transmission Element with a Fiber Bragg Grating Sensor Embedded in a Replaceable Measuring Head
Previous Article in Special Issue
Selecting Some Variables to Update-Based Algorithm for Solving Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Coverage Optimization in a WSN Based on an Improved COOT Bird Algorithm

1
Electrical Engineering College, Guizhou University, Guiyang 550025, China
2
Power China Guizhou Electric Power Engineering Co., Ltd., Guiyang 550025, China
3
Power China Guizhou Engineering Co., Ltd., Guiyang 550001, China
4
Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3383; https://doi.org/10.3390/s22093383
Submission received: 9 April 2022 / Revised: 22 April 2022 / Accepted: 25 April 2022 / Published: 28 April 2022
(This article belongs to the Special Issue Nature-Inspired Algorithms for Sensor Networks and Image Processing)

Abstract

:
To address the problems of uneven distribution and low coverage of wireless sensor network (WSN) nodes in random deployment, a node coverage optimization strategy with an improved COOT bird algorithm (COOTCLCO) is proposed. Firstly, the chaotic tent map is used to initialize the population, increase the diversity of the population, and lay the foundation for the global search for the optimal solutions. Secondly, the Lévy flight strategy is used to perturb the individual positions to improve the search range of the population. Thirdly, Cauchy mutation and an opposition-based learning strategy are fused to perturb the optimal solutions to generate new solutions and enhance the ability of the algorithm to jump out of the local optimum. Finally, the COOTCLCO algorithm is applied to WSN coverage optimization problems. Simulation results show that COOTCLCO has a faster convergence speed and better search accuracy than several other typical algorithms on 23 benchmark test functions; meanwhile, the coverage rate of the COOTCLCO algorithm is increased by 9.654%, 13.888%, 6.188%, 5.39%, 1.31%, and 2.012% compared to particle swarm optimization (PSO), butterfly optimization algorithm (BOA), seagull optimization algorithm (SOA), whale optimization algorithm (WOA), Harris hawks optimization (HHO), and bald eagle search (BES), respectively. This means that in terms of coverage optimization effect, COOTCLCO can obtain a higher coverage rate compared to these algorithms. The experimental results demonstrate that COOTCLCO can effectively improve the coverage rate of sensor nodes and improve the distribution of nodes in WSN coverage optimization problems.

1. Introduction

1.1. Background of Problem

Wireless sensor networks (WSNs) are composed of a large number of low-power sensor nodes with communication functions, and have been widely used in military, industrial, and agricultural control, urban management, biomedicine, environmental detection, disaster relief, and other fields [1,2]. The coverage problem is one of the most fundamental problems in wireless sensor networks, and coverage is an important indicator for evaluating coverage optimization strategies, which has a great impact on the quality of service of wireless sensor networks because it directly determines the monitoring capability of the target monitoring area in the wireless sensor network. Rational and effective deployment of sensor nodes not only minimizes the network cost, but also reduces energy consumption in the sensor power optimization problem [3,4,5,6]. In coverage applications of wireless sensor networks, in order to improve coverage efficiency, all aim to deploy a minimum number of sensor nodes to monitor a specific target area of interest. Generally, sensor nodes are randomly deployed in the target monitoring area, resulting in uneven distribution of sensor nodes and, thus, bringing low coverage. Therefore, it is of great importance to improve the coverage of wireless sensor networks in the monitoring area by deploying sensor nodes rationally and efficiently [7,8].

1.2. Related Works

The node deployment problems in wireless sensor networks can be solved by building an integer linear programming model and then using the methods for solving it. In general, the branch-and-bound method is commonly used to solve the integer programming problem [9,10,11]. For small-scale node deployment problems, integer linear programming methods can be used to solve them. However, for large-scale sensor node deployment problems, the rational and efficient deployment of wireless sensor nodes has been proven to be an NP-hard problem, and finding the optimal solution for such problems remains a challenge. In this background, scholars have proposed the use of metaheuristic algorithms as a solution. Metaheuristic algorithms are approximate optimization algorithms with solutions that escape from local optima, and are widely regarded as effective methods for solving high-dimensional optimization problems, as well as various complex engineering problems. Metaheuristic algorithms can find near-optimal solutions in a reasonable time using limited computational resources, providing a very effective approach to the coverage optimization problems for wireless sensor networks [12,13,14]. Metaheuristic algorithms are classified into four categories of algorithms: evolutionary concepts, animal behavior, physical phenomena, and human behavior [15,16,17]. The first category is based on evolutionary concepts, including genetic algorithms (GAs) [18], genetic programming (GP) [19], evolutionary programming (EP) [20], differential evolution (DE) [21], and biogeography-based optimizers (BBOs) [22]. The second category is based on animal behavior, such as the coot optimization algorithm (COOT) [17], particle swarm optimization (PSO) [23], grey wolf optimizer, GWO) [24], salp swarm algorithm (SSA) [25], butterfly optimization algorithm (BOA) [26], seagull optimization algorithm (SOA) [27], whale optimization algorithm (WOA) [28], Harris hawks optimization (HHO) [29], and bald eagle search (BES) [30]. The third category consists of algorithms based on physical phenomena in nature, such as simulated annealing (SA) [31], black hole algorithm (BH) [32], sine–cosine algorithm (SCA) [33], and ray optimization (RO) [34]. The fourth category is human-behavior-based algorithms, such as teaching–learning-based optimization (TLBO) [35], harmony search (HS) [36], exchange market algorithm (EMA) [37], imperialist competitive algorithm (ICA) [38], and political optimizer (PO) [39].
There are some studies on using metaheuristic algorithms for WSN coverage optimization problems. In [40], ZainEldin et al. proposed a dynamic deployment strategy based on IDDT-GA to maximize the area coverage rate with the minimum number of sensor nodes. However, the proposed algorithm tends to fall into local optima, which affects the optimization of coverage, and there are too few compared algorithms for it to be convincing enough. In [41], Zhang et al. proposed an SA-GWO algorithm for the problem of high aggregation and low coverage rate when sensor nodes are deployed randomly. Although a better coverage effect was achieved, the time complexity of the algorithm was high. In [42], Liu et al. used the ALO algorithm to address the problems of uneven node distribution and incomplete coverage in node deployment. However, the coverage optimization of this algorithm was not ideal, and there was still an uneven distribution of sensor nodes. In [43], Liu et al. proposed an EFWA algorithm to solve the dynamic deployment problem of mobile sensor networks. However, the algorithm would easily fall into a local optimum too early, resulting in a lower coverage rate. In [44], Liao et al. used the GSO algorithm to improve the coverage after random deployment. However, there were obvious coverage holes and node redundancy. In [45], Ozturk et al. used the ABC algorithm to address the dynamic deployment problem in mobile and fixed-sensor scenarios, but there were too few compared algorithms, and the optimized experimental results still had node redundancy. In [46], Zhu et al. proposed an improved hybrid strategy weed algorithm to solve coverage optimization problems of WSNs. However, the problem of coverage holes still existed. It can be seen that it is of value and significance to use metaheuristic algorithms or improved metaheuristic algorithms for node coverage optimization problems. At the same time, the common problems of metaheuristic algorithm, such as slow convergence speed and susceptibility to falling into local optima, will lead to poor coverage optimization effects, which is also the focus of improvement in this paper.
The COOT optimization algorithm [17]—a novel metaheuristic algorithm proposed by the Iranian scholar Iraj Naruei in 2021—has been rapidly applied in the field of engineering optimization. In [47], Memarzadeh et al. used the COOT algorithm to optimize the parameters of a wind power prediction model. In [48], Gouda et al. used the COOT optimization algorithm to test the performance of solar power generator units. In [49], Mahdy et al. combined the COOT optimization algorithm with an anti-windup approach as a way to improve the transient stability of a wave energy conversion system (WECS). In [50], Houssein et al. proposed an improved COOT optimization algorithm for identifying the optimal lithium-ion (Li-ion) battery model parameters. In [51], Alqahtani et al. proposed a COOT-CMO algorithm for selecting the appropriate optimal candidate terms in the automatic query expansion process.
Although the COOT algorithm has been successfully applied and solved some engineering optimization problems, it also has problems, such as sluggish convergence speed and susceptibility to falling into local optima, the same as other metaheuristic algorithms. Therefore, in order to improve the flaws in the original algorithm, four effective strategies—namely, chaotic tent map, Lévy flight, Cauchy mutation, and opposition-based learning—are introduced to the algorithm to propose an improved COOT optimization algorithm, named COOTCLCO.

1.3. Contributions

The main contributions of this paper are as follows:
  • A new and improved algorithm named COOTCLCO, based on the COOT bird algorithm, is proposed.
  • Population diversity is improved by introducing the chaotic tent map to initialize populations. Expanding the search range of the populations by introducing the Lévy flight strategy, the capability of the algorithm to jump out of the local optimum is enhanced by introducing the Cauchy mutation and the opposition-based learning strategy.
  • The optimization capability of the proposed algorithm is tested on unimodal, multimodal, and fixed-dimension multimodal benchmark test functions.
  • The proposed algorithm is compared with seven metaheuristic algorithms in numerical analysis and convergence curves for the performance of finding the best optimal value.
  • An integer linear programming model is used to describe the coverage optimization problem of wireless sensor networks, and the proposed algorithm is used to solve this optimization problem. The proposed algorithm is compared with six metaheuristic algorithms in the coverage optimization problem.

1.4. Notations

The following Table 1 illustrates all of the notations that appear in this paper:

1.5. Organization

The remainder of this paper is organized as follows: Section 2 introduces the node coverage model of WSN. Section 3 describes the basic principles of the COOT optimization algorithm in detail. Section 4 details the improvement strategies of the COOT optimization algorithm. Section 5 introduces the coverage optimization strategy. Section 6 details the experimental design scheme, benchmark test function search performance comparison, coverage optimization performance comparison, and the practical application of COOTCLCO in addressing WSN coverage optimization problems. Section 7 gives a summary.

2. WSN Node Coverage Model

Suppose that q sensor nodes are randomly deployed in a two-dimensional WSN monitoring area with an area of M × N m2, where the set of nodes can be denoted as S = { S 1 , S 2 , S i , S q } , and the coordinates of each node S i can be denoted as ( x i , y i ) , where i = 1 , 2 , q .
For a two-dimensional WSN monitoring area, the network model is as follows:
(1)
Each sensor node is a homogeneous sensor; that is, it has the same parameters, structure, and communication capabilities.
(2)
Each sensor node has sufficient energy, normal communication function, and timely access to data information.
(3)
Each sensor node can move freely, and can update the location information in time.
(4)
The sensing radius of each sensor node is R s and the communication radius is R c , both in units of meters, and R c 2 R s .
The sensing range of a sensor node is a circular area, with the node itself as the center and the sensing radius R s as the radius. Assuming that there are n target monitoring points in this two-dimensional WSN monitoring area, the set of target monitoring points can be denoted as T = { T 1 , T 2 , T j , T n } , and the location coordinates of each target point T j to be monitored are ( x j , y j ) , where j = 1 , 2 , n . If the distance between the target monitoring point T j and any of the sensor nodes is less than or equal to the sensing radius R s , then it can be concluded that T j is covered by the sensor nodes. The Euclidean distance between sensor node S i and target monitoring point T j is defined as:
d ( S i , T j ) = ( x i x j ) 2 + ( y i y j ) 2
The node-sensing model in this paper is a Boolean sensing model; that is, when the sensing radius R s is greater than or equal to d ( S i , T j ) , the probability that the target is monitored is 1; otherwise, the probability that the target is monitored is 0. If the probability that the target point T j to be monitored is covered by the sensor node S i be p , then
p ( S i , T j ) = { 1 R s d ( S i , T j ) 0 R s < d ( S i , T j )
In this two-dimensional WSN monitoring area, the sensor nodes can work cooperatively with one another; that is, any target monitoring point can be covered by more than one sensor at the same time, so the probability that any monitoring target point T j is jointly sensed is:
P ( S , T j ) = 1 i = 1 q ( 1 p ( S i , T j ) )
The coverage rate is defined as the rate of the coverage area of all sensor nodes in the monitoring area to the total area of the monitoring area; thus, the coverage rate of this 2D WSN monitoring area is:
C o v = j = 1 n P ( S , T j ) M × N
Based on the above analysis, the node coverage optimization problem for wireless sensor networks can be described by the following integer linear programming model:
M a x   C o v = j = 1 n P ( S , T j ) M × N
s . t . { j = 1 n P ( S , T j ) 0 1 j n j = 1 n P ( S , T j ) M × N 1 j n d ( S i , T j ) R s 1 i q , 1 j n  
where C o v denotes the objective function for which the maximum coverage rate is required to be solved, S i denotes the i-th sensor node, T j denotes the j-th target point to be monitored, and M × N denotes the size of the monitoring area. The first constraint represents the probability constraint that any monitoring target point T j is jointly sensed. The second constraint indicates that the area covered by all sensor nodes in the monitoring area should be less than the total area of the monitoring area. The third constraint indicates that the Euclidean distance between the sensor node S i and the target monitoring point T j should be less than the sensing radius R s to effectively cover the target monitoring point.
When the size of the sensor nodes to be deployed is relatively large, it takes a lot of time to solve for the coverage problem using integer linear programming methods to obtain the optimal solution. To solve this puzzle effectively, a metaheuristic algorithm is appropriate, because metaheuristic algorithms can give satisfactory results in a tolerable time. Therefore, in this paper, an improved coot bird algorithm is proposed to solve the coverage optimization problem of wireless sensor networks.

3. COOT Optimization Algorithm

The principle of the COOT optimization algorithm is based on the different movement behaviors of coot flocks on the water surface. Coots are small waterbirds that have many different group behaviors on the water surface, with the ultimate goal of the behavior being to move toward food or a specific location. On the water surface, the coot group mainly has four different movement behaviors: random movement, chain movement, adjusting position according to the leader, and leader movement [17]. The process of implementing the COOT algorithm is composed of these four movement behaviors. The specific procedure of the algorithm is as follows [17]:
Initialize the population—random initialization of the population according to Equation (7):
C o o t P o s ( i ) = r a n d ( 1 , d ) × ( u b l b ) + l b
where C o o t P o s ( i ) is the position of the i-th coot, d is the number of variables or problem dimensions, u b is the upper bound of the search space, and l b is the lower bound of the search space. u b and l b are defined as follows:
u b = [ u b 1 , u b 2 , u b d ] , l b = [ l b 1 , l b 2 , l b d ]
After initializing the population, the position of the coot is updated according to the following four movement behaviors.

3.1. Random Movement

In this movement, a position Q is first initialized randomly using Equation (9):
Q = r a n d ( 1 , d ) × ( u b l b ) + l b
In order to avoid getting trapped in a local optimum, the position is updated according to Equation (10):
C o o t P o s ( i ) = C o o t P o s ( i ) + A × R 2 × ( Q C o o t P o s ( i ) )
where R 2 is a random number in the interval [0, 1], and A is determined from Equation (11):
A = 1 L × ( 1 I t e r )
where I t e r is the maximum number of iterations and L is the current number of iterations.

3.2. Chain Movement

The chain movement can be implemented by using the average position of the two coot birds, using Equation (12) to calculate the average position of the two coot birds:
C o o t P o s ( i ) = C o o t P o s ( i 1 ) + C o o t P o s ( i ) 2
where C o o t P o s ( i 1 ) is the location of the second coot bird.

3.3. Adjusting Position According to the Leader

The coot bird updates its own position according to the position of the leader in the group; that is, the coot bird follower in each group moves towards the leader. The leader is selected according to Equation (13):
K = 1 + ( i   M O D   N L )
where K is the number of the leader, i is the number of the coot bird follower, and N L is the number of leaders.
In this movement, the coot bird updates its position according to Equation (14):
C o o t P o s ( i ) = L e a d e r P o s ( k ) + 2 × R 1 × cos ( 2 R π ) × ( L e a d e r P o s ( k ) C o o t P o s ( i ) )
where C o o t P o s ( i ) is the current position of the coot bird, L e a d e r P o s ( k ) is the position of the selected leader, R 1 is a random number in the interval [0, 1], and R is a random number in the interval [−1, 1].

3.4. Leader Movement

In order to find the optimal position, the leader must jump from the existing local optimal position to the global optimal position, using Equation (15) to complete the leader position update:
L e a d e r P o s ( i ) = { B × R 3 × cos ( 2 π R ) × ( g B e s t L e a d e r P o s ( i ) ) + g B e s t R 4 < 0.5 B × R 3 × cos ( 2 π R ) × ( g B e s t L e a d e r P o s ( i ) ) g B e s t R 4 0.5
where g B e s t is the best position that can be found, R 3 and R 4 are the random numbers between the interval [0, 1], and R is the random number between the interval [−1, 1]. B is determined from Equation (16):
B = 2 L × ( 1 I t e r )

4. Improved COOT Optimization Algorithm

The basic COOT algorithm uses random initialization in the initialization process, which reduces the diversity of the initial population which, in turn, affects the performance of the algorithm. Due to the limitation of the search principle of the COOT algorithm, as the number of iterations increases, the individuals in the coot population gradually move closer to the leader with better fitness in the population, and if the leader cannot jump out of the local optimum in time, the whole population will easily fall into the state of local optimum, which reduces the algorithm’s search accuracy. Therefore, this paper proposes an improved COOT optimization algorithm: in the initial stage of the algorithm, a chaotic tent map is added to improve the diversity of the population and lay the foundation for improving the global search ability; subsequently, during the iterative process, the Lévy flight strategy is used to perturb the location of coot individuals to improve the search range of the population and reduce the phenomenon of falling into a local optimum; finally, at the optimal solution location, the Cauchy mutation and the opposition-based learning strategy are fused to perturb the mutation and generate a new solution to further enhance the capability of the algorithm to jump out of the local optimum, while improving the search accuracy of the algorithm.

4.1. Chaotic Tent Map Initializes the Population

Chaotic maps have the characteristics of ergodicity, randomness, and orderliness. Using chaotic variables for optimization searching can improve the diversity of populations and enable the algorithm to jump out of the local optimum, while improving the global search capability. The most common chaotic maps are tent maps, logistic maps, etc. Shan [52] demonstrated that tent maps have a faster search speed in combination with search algorithms compared to logistic maps. Li [53] also demonstrated the effectiveness of tent maps on a swarm intelligence algorithm to enhance population diversity. Therefore, in this paper, a tent map was selected to initialize the population. The expression for the chaotic tent map is defined as follows:
z k + 1 = { z k α 0 < z k α 1 z k 1 α α < z k 1
The chaotic tent map is more effective when α is taken as 0.5, and the distribution of the sequence is more uniform at this time.
Therefore, the expression of the chaotic tent map in this paper is:
z k + 1 = { 2 z k 0 < z k 0.5 2 ( 1 z k ) 0.5 < z k 1
It can also be abbreviated as [54]:
z k + 1 = μ min { z k , 1 z k }
where μ is the parameter that controls the chaotic tent map, and here μ is taken as 2. According to the value of the parameter μ , the bifurcation diagram and Lyapunov exponential curve of the chaotic tent map were drawn, as shown in Figure 1. From Figure 1a, it can be seen that the chaotic tent map starts to bifurcate when the control parameter μ > 1 , and produces an approximately uniformly distributed chaotic sequence when μ = 2 ; as can be seen from Figure 1b, when μ > 1 , that is, the Lyapunov exponent is greater than 0—it indicates that the chaotic tent map shows chaotic phenomena.

4.2. Lévy Flight Strategy

Lévy flight is a type of search for random walks obeying the Lévy distribution, characterized by the occurrence of long-range jumps as a class of non-Gaussian stochastic processes with Markovian properties, whose principle is derived from a probability distribution proposed by the French mathematician Paul Lévy [55]. The simulation of Lévy flight is shown in Figure 2.
Lévy flight involves a large jump in the search process, for which this paper introduces the Lévy flight mechanism to perturb the position update formula of the COOT algorithm. This can effectively boost the diversity of the population, expand the search range, improve the search capability of a single coot bird, and make the algorithm more easily jump out of the local optimum. Lévy flight can be described by the following mathematical Equation [56]:
L e v y ( s , γ , μ ) { γ 2 π exp [ γ 2 ( s u ) ] 1 ( s u ) 3 2 0 < μ < s < ,   γ > 0 0 s 0
where μ is the displacement parameter and γ is the scale parameter.
After Lévy flight is introduced, the position update Equation is [57]:
x i t + 1 = x i t + α L e v y ( λ )
where x i t + 1 denotes the position after the Lévy flight perturbation; x i t denotes the current position α denotes the step control factor; denotes the dot product; and L e v y ( λ ) denotes the random search path, indicating that it obeys the Lévy distribution with parameter λ [57]:
L e v y ( λ ) ~ u = t λ ,   1 < λ 3
Computation of the Lévy flight random search path using the Mantegna algorithm is as follows [58]:
s = μ | v | 1 β ,   0 < β < 2
where β = λ 1 , and β usually takes a value of 1.5. μ and v both obey a normal distribution [59]:
{ μ ~ N ( 0 , δ u 2 ) v ~ N ( 0 , δ v 2 )
{ δ u = [ Γ ( 1 + β ) × sin ( π β 2 ) Γ ( 1 + β 2 ) × β × 2 β 1 2 ] 1 β δ v = 1
Now, using the position update idea of Equation (21), the position update Equation of Equations (10), (12) and (14) are improved to obtain the following new position update Equations:
C o o t P o s ( i ) = C o o t P o s ( i ) + A × R 2 × ( Q C o o t P o s ( i ) ) × L e v y
C o o t P o s ( i ) = C o o t P o s ( i 1 ) + C o o t P o s ( i ) 2 × L e v y
C o o t P o s ( i ) = L e a d e r P o s ( k ) + 2 × R 1 × cos ( 2 R π ) ×        ( L e a d e r P o s ( k ) C o o t P o s ( i ) ) × L e v y

4.3. Fusing Cauchy Mutation and Opposition-Based Learning

Opposition-based learning [60] was proposed by Tizhoosh in 2005, and its primary idea is to take the current solution to a problem and then find its corresponding reverse solution through the opposition-based learning mechanism, and evaluate the original solution and the reverse solution to finally retain the optimal solution.
If x [ a , b ] , then the distance from x to a is | x a | , then the opposition-based learning number x of x can be defined as follows [60]:
x = a + b x
Then, the distance from x to b is | b x | = | b ( a + b x ) | = | x a | , and these two distances are equal. The relationship between any real number and its opposition-based learning number is shown in Figure 3.
If x = ( x 1 , x 2 , , x n ) is a point in n-dimensional space, where x i [ a i , b i ] , the corresponding reverse point is [60]:
x = ( x 1 , x 2 , , x n )
Its corresponding reverse solution is [60]:
x i = a i + b i x i
The basic COOT algorithm needs to select the optimal position by calculating the fitness after each iteration, and if its position update formula is not perturbed, the algorithm is prone to falling into a local optimum. In the previous section, this paper introduced Lévy flight to perturb its position update formula, but to further enhance the algorithm’s ability to jump out of local optima, this subsection combines Cauchy mutation and the opposition-based learning strategy to perform position update for the movement of adjusting position according to the leader. This step of the movement is where the coot bird updates its own position according to the position of the leader in the group, and it is important to choose the best leader for this phase. Therefore, we find its corresponding reverse solution according to the chosen leader position, which can provide more opportunities to find potential optimal solutions, further boost the diversity of the population based on the Lévy flight perturbation, enhance the global search capability of the algorithm, and prevent the algorithm from falling into a local optimum. Now, the idea of opposition-based learning and the COOT algorithm are combined, and Equation (31) is used to improve Equation (14), which leads to the reverse solution of the leader position and the position update equation based on opposition-based learning:
X L e a d e r P o s ( i ) = u b + r ( l b X L e a d e r P o s ( i ) )
X G b e s t N e w ( i + 1 ) = X L e a d e r P o s ( i ) + b 1 2 × R 1 cos ( 2 π R ) ( X L e a d e r P o s ( i ) X C o o t P o s ( i ) )
where X L e a d e r P o s ( i ) in Equation (32) is the inverse solution of the current leader position at the i-th iteration, u b and l b denote upper and lower bounds, r denotes the random number matrix, and denotes the dot product. X G b e s t N e w ( i + 1 ) in Equation (33) denotes the latest position of the i + 1-th iteration, R 1 is a random number between the interval [0, 1], R is a random number between the interval [−1, 1], X C o o t P o s ( i ) denotes the position of the coot bird follower of the i-th iteration, and b 1 denotes the information exchange control parameter, which is calculated as follows [61]:
b 1 = ( 1 i m a x _ I t e r ) i
The Cauchy distribution is one of the common continuous-type distributions in probability theory, and its one-dimensional probability density function expression can be defined as follows [62]:
f ( x ) = 1 π T ( x 2 + T 2 ) ,   - < x < , T > 0
where T denotes the control parameter. The corresponding distribution function is [62]:
F ( x ) = 1 2 + 1 π arctan ( x T )
For Equation (35), when T=1, the standard Cauchy distribution probability density function [62] is obtained as shown in Equation (37):
f ( x ) = 1 π 1 ( x 2 + 1 ) ,   - < x <
A comparison of the probability density function curves of the standard Cauchy distribution and the Gaussian distribution [63] is shown in Figure 4. From the figure, it can be seen that the standard Cauchy distribution is similar to the standard Gaussian distribution, in that both of them are continuous probability distributions. However, the Cauchy distribution has a long and flat shape at both ends, approaches zero at a slower rate, and has a smaller peak near the origin compared to the Gaussian distribution. Therefore, the Cauchy distribution has a wider distribution range, and allows for greater mutation than the Gaussian distribution.
Introducing the Cauchy mutation into the position update formula of the COOT algorithm and exploiting the perturbation ability of the Cauchy mutation operator will further improve the diversity of the population, enhance the global search ability of the algorithm, and prevent the algorithm from falling into a local optimum. The new position update equation is generated using the Cauchy mutation as follows:
X G b e s t N e w ( i + 1 ) = X C o o t P o s ( i ) + C a u c h y ( 0 , 1 ) X C o o t P o s ( i )
where X G b e s t N e w ( i + 1 ) denotes the latest position of the i + 1-th iteration after perturbation by the Cauchy mutation, X C o o t P o s ( i ) denotes the position of the coot bird of the i-th iteration, and C a u c h y ( 0 , 1 ) denotes the standard Cauchy distribution.
In order to enhance the performance of the algorithm for finding the best solution, the Cauchy mutation and the opposition-based learning strategy can be fused, and the selection probability Ps for deciding which strategy to choose for the position update can be defined using a dynamic selection mechanism, so that both of them are executed alternately with a certain probability as follows [61]:
P s = exp ( 1 i m a x _ I t e r ) 20 + η
where η is the control parameter, and generally takes the value of 0.05 [61].
In this dynamic selection mechanism, if the random number r a n d < P s , the position is updated using the opposition-based learning strategy; otherwise, the position is updated using the Cauchy mutation strategy.

4.4. Implementation Steps of COOTCLCO Algorithm

Step 1: Set the parameters of population size N, maximum number of iterations max_Iter, number of leaders N L e a d e r , number of followers N c o o t , etc., and randomly initialize the positions of coot bird followers and leaders according to Equation (9).
Step 2: Enhance the diversity of the population by using the chaotic tent map in Equation (18).
Step 3: Update the position of the coot bird follower according to the Lévy flight perturbation strategy introduced by Equations (26)–(28).
Step 4: Update the position of the coot bird leader according to Equation (15).
Step 5: Calculate the fitness of the coot bird followers and leaders, and compare them to select the best fitness value.
Step 6: Select the Cauchy mutation or the opposition-based learning strategy according to Equation (39) to perturb the current optimal solution and generate a new solution.
Step 7: Determine whether the end condition is reached; if yes, proceed to the next step; otherwise, return to Step 3.
Step 8: The program ends and outputs the optimal fitness value and the best position.

4.5. COOTCLCO Algorithm Time Complexity Analysis

Suppose that the number of populations of the algorithm is N, the dimension of the search space is D, and the maximum number of iterations is T. Then, for the basic COOT algorithm, its time complexity is O(NDT). For the COOTCLCO algorithm, its time complexity is analyzed as follows:
(1)
The time complexity of initializing the population using the chaotic tent map is O(ND). Thus, the required time complexity is O(NDT) + O(ND) = O(NDT) in the case of introducing only the chaotic tent map.
(2)
The time complexity of perturbing the individual positions using the Lévy flight strategy is O(ND), and the time complexity of the algorithm is O(NDT) after T iterations. Thus, the required time complexity is O(NDT) + O(NDT) = O(NDT) when only the Lévy flight strategy is introduced.
(3)
The time complexity of the algorithm is O(NDT) + O(NDT) after T iterations by fusing the Cauchy mutation and the opposition-based learning strategy and perturbing the optimal solution’s position. Thus, the required time complexity is O(NDT) + O(NDT) + O(NDT) = O(NDT) with the introduction of only the fused Cauchy mutation and the opposition-based learning strategy.
Therefore, after introducing the above three improvement strategies, the time complexity of the COOTCLCO algorithm is O(COOTCLCO) = O(NDT) + O(NDT) + O(NDT) = O(NDT). In summary, the time complexity of the COOTCLCO algorithm is the same as that of the COOT algorithm, thus showing that the improvement strategy proposed in this paper based on the COOT algorithm does not increase the time complexity of the algorithm.

5. Coverage Optimization Strategy

In this paper, the location-seeking process of nodes in the coverage optimization problem is abstracted as the process of making different movement behaviors of the coot bird group toward food or a specific location, and the optimal solution is the target location of each node deployed. The goal of WSN coverage optimization based on the COOTCLCO algorithm is to maximize the coverage of the target monitoring area by using a certain number of sensor nodes and optimizing the locations where they will be deployed. The flowchart of the coverage optimization algorithm is shown in Figure 5. Each coot bird individual in the algorithm represents a coverage distribution, and the specific algorithm steps are as follows:
Step 1: Input parameters such as the number of nodes q, perception radius Rs, area of region M × N, etc., and randomly initialize the positions of the coot bird followers and leaders according to Equation (9).
Step 2: Boost the diversity of the population using the chaotic tent map in Equation (18), and calculate the initial coverage according to Equation (4).
Step 3: Update the position of the coot bird followers according to the Lévy flight perturbation strategy introduced by Equations (26)–(28).
Step 4: Update the position of the coot bird leaders according to Equation (15).
Step 5: Calculate the fitness of the coot bird followers and leaders, and update the coverage rate according to Equation (4), with the coverage rate Cov as the objective function, to find the current best node location.
Step 6: Select the Cauchy mutation or the opposition-based learning strategy according to Equation (39) to perturb the current optimal solution and generate a new solution.
Step 7: Determine whether the end condition is reached; if yes, proceed to the next step; otherwise, return to Step 3.
Step 8: The program ends and the node optimal coverage rate is output.

6. Simulation Experiments and Analysis

6.1. Experimental Design

In order to verify the optimization performance of COOTCLCO and its effectiveness in WSN node coverage optimization, two sets of comparative experiments were designed in this paper: (1) COOTCLCO was compared with the optimization performance of seven optimization algorithms, namely, COOT [17], PSO [23], GWO [24], SSA [25], BOA [26], SOA [27], and SCA [33]; (2) COOTCLCO was compared with six optimization algorithms—PSO, BOA, SOA, WOA [28], HHO [29], and BES [30]—on the WSN node coverage optimization problem.
In this paper, 23 benchmark functions were used to test the algorithm’s performance in finding the optimum, and these 23 benchmark functions can be divided into three categories. Table 2 lists 7 unimodal benchmark functions; Table 3 lists 6 multimodal benchmark functions with multiple local optimal solutions, and the number of local optimal solutions increases exponentially with the number of dimensions; and Table 4 lists 10 fixed-dimension multimodal benchmark functions.
Experimental simulation environment for this paper was as follows: Windows 10 OS, Intel Core i7-10750H CPU @2.60 GHz, NVIDIA GeForce GTX 1650Ti graphics card, SK Hynix 16 GB RAM, MATLAB 2019b simulation platform.

6.2. Performance Comparison on Benchmark Functions

In this section, the optimization ability of COOTCLCO is compared with that of COOT, PSO, GWO, SSA, BOA, SOA, and SCA. For all algorithms, the population size N = 30 and the maximum number of iterations m a x _ I t e r = 1500 . In order to attenuate the chance of the experiment, each algorithm was run 30 times independently for the same test function, taking the average value and standard deviation of the experimental results for comparison, and the parameter settings of the comparison algorithms are detailed in Table 5. In order to test the difference between the COOTCLCO algorithm and the comparison algorithm, the Wilcoxon signed-rank test was used. At the significance level of 0.05, “+”, “≈”, and “−” indicate that the performance of COOTCLCO is superior to, similar to, or inferior to that of the comparison algorithm, respectively. Meanwhile, the average values derived from each algorithm under each test function were ranked. The experimental average value (avg), standard deviation (std), Wilcoxon signed-rank test results (W), and average value ranking (R) are presented in Table 6 and Table 7, and the best average value and standard deviation derived in each test function are shown in bold. Finally, the overall Wilcoxon signed-rank test results, the total average value ranking, and the overall algorithm ranking results are given in Table 8.

6.2.1. Analysis of Numerical Results

The average value, standard deviation, Wilcoxon signed-rank test results, and average value ranking of the simulation experiments for the 8 optimization algorithms on the 23 benchmark test functions are given in Table 5, Table 6 and Table 7. The unimodal test function is suitable for evaluating the exploitation of the algorithms. In terms of average value and standard deviation, for the test functions F1, F2, and F5, GWO has the best performance, while COOTCLCO ranks as the second-best-performing algorithm. For F3 and F4, COOTCLCO is the best-performing optimizer, followed by COOT’s search ability, while SCA is the worst-performing optimization algorithm in these two test functions. For the test functions F6 and F7, COOTCLCO has an average performance in terms of the optimization ability, ranking fourth in the comparison of these eight algorithms tested. It is worth mentioning that in the test function F6, PSO exhibits the strongest optimization ability. Overall, among the unimodal test functions, COOTCLCO is the best overall performing algorithm in terms of optimization ability and stability, and is very effective and competitive with the other seven metaheuristics, while the test results show that COOTCLCO has a good exploitation capability.
Multimodal test functions and fixed-dimension test functions are suitable for assessing the exploration ability of the algorithms. Among the multimodal functions, for the test function F8, the optimization effect of the COOTCLCO algorithm is the best, while SSA and COOT are the second- and third-best-performing algorithms. For F9 and F10, COOTCLCO obtained the best results compared to the other algorithms; in F9, the second- and third-ranked test results were COOT and GWO, respectively; in F10, GWO and BOA were second only to COOTCLCO in terms of finding the best results. For F11, BOA had the best test results, while COOTCLCO followed closely. For F12, COOTCLCO outperformed the other algorithms in terms of average value and standard deviation. For F13, COOTCLCO’s test results ranked second, and it is worth mentioning that PSO performed the best in terms of average value and standard deviation results. Among the fixed-dimension functions, for F14, F20, and F21, COOTCLCO performed better than the other algorithms for both the average value and the standard deviation. For F15, the BOA test results were the best, and its standard deviation results prove that BOA has a strong stability in the test function F15, where the COOTCLCO test results rank second. For F16, F17, and F18, the eight algorithms can generally find the optimal value of the test function in terms of the average value of the optimal value; in terms of the standard deviation, COOTCLCO has the best stability; finally, COOTCLCO wins in the test results of F16, F17, and F18 in terms of stability. For F19, the PSO test results proved it to be the best performing algorithm, while COOTCLCO ranked as the second-best-performing algorithm. For F22 and F23, while COOTCLCO’s test results ranked only third and fourth, respectively, the best-performing GWO’s test results were only marginally stronger than COOTCLCO’s, as the two were very close in their optimization results.
All in all, for the multimodal and fixed-dimension test functions, COOTCLCO ranked first 10 times and second 4 times out of 16 test results, which is evidence that the COOTCLCO algorithm has an extremely strong exploration capability.

6.2.2. Analysis of Convergence Curves

Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12 show the convergence curves of the eight algorithms of COOTCLCO, COOT, PSO, GWO, SSA, BOA, SOA, and SCA in the seven selected benchmark functions F3, F4, F8, F11, F12, F20, and F21, as well as the three-dimensional space diagram of the seven test functions. For all algorithms, the population size N = 30 and the maximum number of iterations m a x _ I t e r = 1500 are set. In order to reduce the contingency of the experiment, the curves given in these figures are the average convergence curves obtained from 30 independent experiments.
Figure 6 illustrates that COOTCLCO has a strong global search capability, and it finds a better global optimal value compared to other algorithms. It can quickly jump out of the local optimum in the middle stage of the algorithm’s iteration so as to continue to find and approach the theoretical optimal value of the function, which benefits from the effect of the Lévy flight mechanism and the algorithm’s improvement by combining the Cauchy mutation and the opposition-based learning. Although the search principle of COOT itself also has the ability to jump out of the local optimum, as can be seen from the figure, this ability is still not powerful enough compared to COOTCLCO. Figure 7 shows that the convergence rates of COOTCLCO, COOT, and GWO are very close in the early stages of the iteration, but both COOT and GWO fall into the local optimum too early, while COOTCLCO continues its search with a faster convergence rate until it obtains an optimal solution close to the theoretical optimal value. Figure 8 shows that PSO, SCA, and BOA fall into a stagnant state prematurely at the early stage of the algorithm’s iteration and, thus, produce bad search results. In contrast, COOTCLCO and COOT keep searching forward in a very stable state, and it can be seen from the figure that the whole process almost never falls into a local optimum, but COOTCLCO eventually has better search accuracy than COOT. Figure 9 shows that COOTCLCO was looking for the optimal value of the target with a very fast convergence speed, and eventually both COOTCLCO and BOA found results close to the theoretical optimal value. However, COOT was stuck in the local optimal state for most of the algorithm’s iteration, and performed poorly overall. Figure 10 shows that SOA, SCA, and SSA were at a standstill at the beginning of the algorithms’ iteration, and although the search started at a very fast convergence rate at the middle stage of the iteration, the algorithms fell into local optima at a later stage. Figure 11 shows that there is almost no difference in the convergence speed and convergence accuracy of the COOTCLCO, COOT, PSO, and GWO algorithms but, relatively speaking, the result of COOTCLCO is closest to the theoretical optimal value. Figure 12 shows that both COOTCLCO and GWO exhibit good search accuracy, but the convergence speed of COOTCLCO is significantly better than that of GWO. In summary, as the test results in Table 5, Table 6 and Table 7 and Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12 show, COOTCLCO achieves very competitive results for most of the benchmark functions, indicating that it has reliable convergence speed as well as better exploration capability.

6.3. Coverage Performance Simulation Experiment and Analysis

To verify the effect of COOTCLCO on WSN node coverage optimization, six optimization algorithms—PSO, BOA, SOA, WOA, HHO, and BES—were selected for comparison on the WSN node coverage optimization problem. The sensor nodes were deployed in a square monitoring area of 100 m × 100 m, the sensing radius of the sensor nodes was Rs = 10 m, the communication radius was Rc = 20 m, the number of sensor nodes was denoted by q, and the number of iterations was denoted by Iteration. The experimental parameters of the node deployment area were set as shown in Table 9. Three sets of comparison experiments were designed in this section: (1) 30 independent experiments with different algorithms for 25, 35, and 45 sensor nodes, to plot their average coverage rate curves; (2) initial coverage diagram and coverage optimization diagram of COOTCLCO for node coverage optimization plotted for 25, 35, and 45 sensor nodes; (3) initial coverage diagram and coverage optimization diagram of COOTCLCO for node coverage optimization plotted for 45 sensor nodes and 500, 1000, and 1500 iterations.

6.3.1. Comparative Experiment 1 and Result Analysis

In the first type of comparative experiments, in order to reduce the contingency of the experiment, 30 independent experiments were conducted with different algorithms for 25, 35, and 45 sensor nodes; the average value of their coverage rate was taken, and their average coverage rate curves were plotted as shown in Figure 13. Table 10 gives the comparison of the average coverage rate results of the seven algorithms, and the comparison of the average coverage rates in the cases of different node numbers is shown in Figure 14. As can be seen from Figure 13, COOTCLCO achieved the best coverage rate in all three cases, and its average final coverage rate was 75.329%, 90.332% and 96.990%, respectively; of course, HHO and BES also obtained a good coverage effect, while BOA had the worst performance in node coverage optimization. In the case of changing only the number of sensor nodes, the coverage rate increased as the number of nodes continued to increase. Overall, COOTCLCO outperformed the other six algorithms in terms of WSN coverage optimization.

6.3.2. Comparative Experiment 2 and Result Analysis

In the second type of comparison experiments, different numbers of sensor nodes were randomly deployed in a 100 m × 100 m area with a sensing radius of 10 m and a communication radius of 20 m for each sensor node, and the maximum number of iterations was 1500. The initial coverage diagram and coverage optimization diagram of COOTCLCO for node coverage optimization are shown in Figure 15 for the cases of 25, 35, and 45 sensor nodes. Figure 16 gives a comparison of the coverage rates before and after optimization by COOTCLCO under different numbers of nodes. In Figure 15a, the number of sensor nodes is 25, and the initial coverage ratio of the coverage diagram initialized randomly is 58.81%, while after optimization by COOTCLCO the final coverage rate is obtained as 77.28%, which is an 18.47% increase in coverage. In Figure 15b, the number of sensor nodes is 35, and the initial coverage rate of the coverage diagram initialized randomly is 69.76%, while after optimization by COOTCLCO the final coverage rate is obtained as 94.23%, which is a 24.47% increase in coverage. In Figure 15c, the number of sensor nodes is 45, and the initial coverage rate of the coverage diagram initialized randomly is 79.61%, while after optimization by COOTCLCO the final coverage rate is obtained as 98.07%, which is an 18.46% increase in coverage. It can be seen that in these three cases, after COOTCLCO optimization, the locations of the randomly initialized sensor nodes become neat and orderly rather than haphazard, which effectively improves the coverage rate of the deployment area.

6.3.3. Comparative Experiment 3 and Result Analysis

In the third type of comparison experiments, similarly, the area was set to 100 m × 100 m, the sensing radius of each sensor node was 10 m, and the communication radius was 20 m. We only changed the number of iterations, where the number of sensor nodes was 45 and the number of iterations was 500, 1000, or 1500; the initial coverage diagram of the nodes and the coverage diagram optimized by COOTCLCO are given in Figure 17. Figure 18 gives a comparison of the coverage rates before and after optimization by COOTCLCO when the number of nodes is 45. In Figure 17a, the number of iterations is set to 500, and the initial coverage rate of the coverage diagram initialized randomly is 79.95%, while after optimization by COOTCLCO the final coverage is obtained as 91.28%, which is an 11.33% increase in coverage. In Figure 17b, the number of iterations is set to 1000, and the initial coverage rate of the coverage diagram initialized randomly is 81.08%, while after optimization by COOTCLCO the final coverage is obtained as 95.86%, which is a 14.78% increase in coverage. In Figure 17c, the number of iterations is set to 1500, and the initial coverage rate of the coverage diagram initialized randomly is 79.61%, while after optimization by COOTCLCO the final coverage is obtained as 98.07%, which is an 18.46% increase in coverage. It can be seen that the final coverage rate increased by only 6.79% when the number of iterations was increased from 500 to 1500, while in the second group of comparison experiments, the final coverage rate improved by 20.79% when the number of sensor nodes was increased from 25 to 45. Therefore, in the case of changing only the number of sensor nodes or only the number of iterations, reasonably and finitely increasing the number of sensor nodes is more likely to greatly improve the coverage rate of the target monitoring area compared to increasing the number of iterations.

7. Conclusions

Aiming at the problems of uneven node distribution and low coverage of the target monitoring area when randomly deploying sensor nodes in WSNs, a COOTCLCO algorithm for node coverage optimization in WSNs is proposed in the paper. COOTCLCO uses a chaotic tent map to initialize the population based on the original COOT algorithm, which increases the diversity of the population and enhances the traversal of the search space by the COOT population. The Lévy flight strategy is introduced to perturb individual positions, which can expand the search range of the population and reduce the possibility of the algorithm falling into a local optimum. The algorithm then combines the Cauchy mutation and the opposition-based learning strategy to perturb the optimal solution positions and generate new solutions, which further enhances the ability of the algorithm to jump out of the local optimum. In order to verify the optimization performance of COOTCLCO, 23 benchmark functions were used to test the optimization performance of the algorithm, which was compared with seven other optimization algorithms: COOT, PSO, GWO, SSA, BOA, SOA and SCA. By analyzing the numerical results and convergence curves of the simulation experiments, we found that COOTCLCO has reliable convergence speed as well as better global exploration capability. To verify the capability of COOTCLCO on the WSN node coverage optimization problem, we compared it with six optimization algorithms, namely, PSO, BOA, SOA, WOA, HHO, and BES. The experimental results show that COOTCLCO obtained the highest average coverage rate under the same test conditions, and the coverage rate convergence curves indicate that COOTCLCO can improve the coverage rate of WSN nodes quickly and effectively. This means that COOTCLCO only requires the least number of sensor nodes to achieve the same coverage rate in the same target monitoring area, which reduces the deployment cost of sensor nodes.
However, the work of this paper has some limitations. Although our proposed algorithm has a significant effect on the improvement of the coverage rate, there are still shortcomings in the experimental design. Firstly, our experiments were designed to optimize the coverage rate in an ideal environment, while in a real complex environment the influence of obstacles should be taken into account. Secondly, this paper only considers the coverage optimization of a two-dimensional plane, and the coverage optimization of wireless sensor networks can be extended to three-dimensional space. Thirdly, the coverage optimization objective of this paper is relatively singular, and only the coverage rate optimization is considered, while the other coverage optimization indices of wireless sensor networks such as network energy consumption, network life cycle, network security, and sensor power optimization should also be taken into consideration in the real environment. Therefore, for the next step, we will continue to improve and refine the algorithm and conduct further research on the above-mentioned limitations.

Author Contributions

Conceptualization, Y.H. and J.Y.; methodology, Y.H. and J.Y.; software, Y.H.; validation, all authors; formal analysis, all authors; investigation, Y.H. and J.Y.; data curation, Y.H.; writing—original draft preparation, Y.H.; writing—review and editing, all authors; visualization, Y.H. and J.Y.; supervision, J.Y., X.L., J.Z. and T.Q.; project administration, all authors; funding acquisition, J.Y., X.L., J.Z., T.Q., W.W. and Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the NNSF of China (No.61640014, No.61963009), the Industrial Project of Guizhou Province (No. Qiankehe Zhicheng [2022]Yiban017, [2019]2152), the Innovation group of Guizhou Education Department (No. Qianjiaohe KY [2021]012), the Science and Technology Fund of Guizhou Province (No. Qiankehejichu [2020]1Y266), the CASE Library of IoT (KCALK201708), and the platform about IoT of Guiyang National High Technology Industry Development Zone (No. 2015).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Akyildiz, I.F.; Su, W.; Sankarasubramaniam, Y.; Cayirci, E. Wireless Sensor Networks: A Survey. Comput. Netw. 2002, 38, 393–422. [Google Scholar] [CrossRef] [Green Version]
  2. Zhu, C.; Zheng, C.; Shu, L.; Han, G. A Survey on Coverage and Connectivity Issues in Wireless Sensor Networks. J. Netw. Comput. Appl. 2012, 35, 619–632. [Google Scholar] [CrossRef]
  3. Miao, Z.; Yuan, X.; Zhou, F.; Qiu, X.; Song, Y.; Chen, K. Grey Wolf Optimizer with an Enhanced Hierarchy and Its Application to the Wireless Sensor Network Coverage Optimization Problem. Appl. Soft Comput. 2020, 96, 106602. [Google Scholar] [CrossRef]
  4. Rebai, M.; Le Berre, M.; Snoussi, H.; Hnaien, F.; Khoukhi, L. Sensor Deployment Optimization Methods to Achieve Both Coverage and Connectivity in Wireless Sensor Networks. Comput. Oper. Res. 2015, 59, 11–21. [Google Scholar] [CrossRef]
  5. Tariq, N.; Asim, M.; Maamar, Z.; Farooqi, M.Z.; Faci, N.; Baker, T. A Mobile Code-Driven Trust Mechanism for Detecting Internal Attacks in Sensor Node-Powered IoT. J. Parallel Distrib. Comput. 2019, 134, 198–206. [Google Scholar] [CrossRef]
  6. Tariq, N.; Asim, M.; Khan, F.A.; Baker, T.; Khalid, U.; Derhab, A. A Blockchain-Based Multi-Mobile Code-Driven Trust Mechanism for Detecting Internal Attacks in Internet of Things. Sensors 2021, 21, 23. [Google Scholar] [CrossRef]
  7. Deepa, R.; Venkataraman, R. Enhancing Whale Optimization Algorithm with Levy Flight for Coverage Optimization in Wireless Sensor Networks. Comput. Electr. Eng. 2021, 94, 107359. [Google Scholar] [CrossRef]
  8. Wang, S.; Yang, X.; Wang, X.; Qian, Z. A Virtual Force Algorithm-Lévy-Embedded Grey Wolf Optimization Algorithm for Wireless Sensor Network Coverage Optimization. Sensors 2019, 19, 2735. [Google Scholar] [CrossRef] [Green Version]
  9. Zhang, T.; Tao, X.; Cui, Q. Joint Multi-Cell Resource Allocation Using Pure Binary-Integer Programming for LTE Uplink. In Proceedings of the 2014 IEEE 79th Vehicular Technology Conference (VTC Spring), Seoul, Korea, 18–21 May 2014; pp. 1–5. [Google Scholar]
  10. Liang, D.; Shen, H.; Chen, L. Maximum Target Coverage Problem in Mobile Wireless Sensor Networks. Sensors 2020, 21, 184. [Google Scholar] [CrossRef]
  11. Zou, Y.; Chakrabarty, K. A Distributed Coverage- and Connectivity-Centric Technique for Selecting Active Nodes in Wireless Sensor Networks. IEEE Trans. Comput. 2005, 54, 978–991. [Google Scholar] [CrossRef]
  12. Tsai, C.-W.; Tsai, P.-W.; Pan, J.-S.; Chao, H.-C. Metaheuristics for the Deployment Problem of WSN: A Review. Microprocess. Microsyst. 2015, 39, 1305–1317. [Google Scholar] [CrossRef]
  13. Alia, O.M.; Al-Ajouri, A. Maximizing Wireless Sensor Network Coverage With Minimum Cost Using Harmony Search Algorithm. IEEE Sens. J. 2017, 17, 882–896. [Google Scholar] [CrossRef]
  14. Senouci, M.R.; Abdellaoui, A. Efficient Sensor Placement Heuristics. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–6. [Google Scholar]
  15. Mahdavi, S.; Shiri, M.E.; Rahnamayan, S. Metaheuristics in Large-Scale Global Continues Optimization: A Survey. Inf. Sci. 2015, 295, 407–428. [Google Scholar] [CrossRef]
  16. Dragoi, E.N.; Dafinescu, V. Review of Metaheuristics Inspired from the Animal Kingdom. Mathematics 2021, 9, 2335. [Google Scholar] [CrossRef]
  17. Naruei, I.; Keynia, F. A New Optimization Method Based on COOT Bird Natural Life Model. Expert Syst. Appl. 2021, 183, 115352. [Google Scholar] [CrossRef]
  18. Holland, J.H. Genetic Algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  19. Koza, J.R. Genetic Programming as a Means for Programming Computers by Natural Selection. Stat. Comput. 1994, 4, 87–112. [Google Scholar] [CrossRef]
  20. Yao, X.; Liu, Y.; Lin, G. Evolutionary Programming Made Faster. IEEE Trans. Evol. Comput. 1999, 3, 82–102. [Google Scholar]
  21. Storn, R. On the Usage of Differential Evolution for Function Optimization. In Proceedings of the North American Fuzzy Information Processing, Berkeley, CA, USA, 19–22 June 1996; pp. 519–523. [Google Scholar]
  22. Simon, D. Biogeography-Based Optimization. IEEE Trans. Evol. Comput. 2008, 12, 702–713. [Google Scholar] [CrossRef] [Green Version]
  23. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995. [Google Scholar]
  24. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  25. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A Bio-Inspired Optimizer for Engineering Design Problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  26. Arora, S.; Singh, S. Butterfly Optimization Algorithm: A Novel Approach for Global Optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  27. Dhiman, G.; Kumar, V. Seagull Optimization Algorithm: Theory and Its Applications for Large-Scale Industrial Engineering Problems. Knowl.-Based Syst. 2019, 165, 169–196. [Google Scholar] [CrossRef]
  28. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  29. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris Hawks Optimization: Algorithm and Applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  30. Alsattar, H.A.; Zaidan, A.A.; Zaidan, B.B. Novel Meta-Heuristic Bald Eagle Search Optimisation Algorithm. Artif. Intell. Rev. 2020, 53, 2237–2264. [Google Scholar] [CrossRef]
  31. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by Simulated Annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef]
  32. Hatamlou, A. Black Hole: A New Heuristic Optimization Approach for Data Clustering. Inf. Sci. 2013, 222, 175–184. [Google Scholar] [CrossRef]
  33. Mirjalili, S. SCA: A Sine Cosine Algorithm for Solving Optimization Problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  34. Kaveh, A.; Khayatazad, M. A New Meta-Heuristic Method: Ray Optimization. Comput. Struct. 2012, 112–113, 283–294. [Google Scholar] [CrossRef]
  35. Rao, R.V.; Savsani, V.J.; Vakharia, D.P. Teaching-Learning-Based Optimization: A Novel Method for Constrained Mechanical Design Optimization Problems. Comput.-Aided Des. 2011, 43, 303–315. [Google Scholar] [CrossRef]
  36. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A New Heuristic Optimization Algorithm: Harmony Search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  37. Ghorbani, N.; Babaei, E. Exchange Market Algorithm. Appl. Soft Comput. 2014, 19, 177–187. [Google Scholar] [CrossRef]
  38. Atashpaz-Gargari, E.; Lucas, C. Imperialist Competitive Algorithm: An Algorithm for Optimization Inspired by Imperialistic Competition. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; pp. 4661–4667. [Google Scholar]
  39. Askari, Q.; Younas, I.; Saeed, M. Political Optimizer: A Novel Socio-Inspired Meta-Heuristic for Global Optimization. Knowl.-Based Syst. 2020, 195, 105709. [Google Scholar] [CrossRef]
  40. ZainEldin, H.; Badawy, M.; Elhosseini, M.; Arafat, H.; Abraham, A. An Improved Dynamic Deployment Technique Based-on Genetic Algorithm (IDDT-GA) for Maximizing Coverage in Wireless Sensor Networks. J. Ambient Intell. Humaniz. Comput. 2020, 11, 4177–4194. [Google Scholar] [CrossRef]
  41. Zhang, Y.; Cao, L.; Yue, Y.; Cai, Y.; Hang, B. A Novel Coverage Optimization Strategy Based on Grey Wolf Algorithm Optimized by Simulated Annealing for Wireless Sensor Networks. Comput. Intell. Neurosci. 2021, 2021, 1–14. [Google Scholar] [CrossRef]
  42. Liu, W.; Yang, S.; Sun, S.; Wei, S. A Node Deployment Optimization Method of WSN Based on Ant-Lion Optimization Algorithm. In Proceedings of the 2018 IEEE 4th International Symposium on Wireless Systems within the International Conferences on Intelligent Data Acquisition and Advanced Computing Systems (IDAACS-SWS), Lviv, Ukraine, 20–21 September 2018; pp. 88–92. [Google Scholar]
  43. Liu, X.; Zhang, X.; Zhu, Q. Enhanced Fireworks Algorithm for Dynamic Deployment of Wireless Sensor Networks. In Proceedings of the 2017 2nd International Conference on Frontiers of Sensors Technologies (ICFST), Shenzhen, China, 14–16 April 2017; pp. 161–165. [Google Scholar]
  44. Liao, W.-H.; Kao, Y.; Li, Y.-S. A Sensor Deployment Approach Using Glowworm Swarm Optimization Algorithm in Wireless Sensor Networks. Expert Syst. Appl. 2011, 38, 12180–12188. [Google Scholar] [CrossRef]
  45. Ozturk, C.; Karaboga, D.; Gorkemli, B. Probabilistic Dynamic Deployment of Wireless Sensor Networks by Artificial Bee Colony Algorithm. Sensors 2011, 11, 6056–6065. [Google Scholar] [CrossRef] [Green Version]
  46. Zhu, F.; Wang, W. A Coverage Optimization Method for WSNs Based on the Improved Weed Algorithm. Sensors 2021, 21, 5869. [Google Scholar] [CrossRef]
  47. Memarzadeh, G.; Keynia, F. A New Optimal Energy Storage System Model for Wind Power Producers Based on Long Short Term Memory and Coot Bird Search Algorithm. J. Energy Storage 2021, 44, 103401. [Google Scholar] [CrossRef]
  48. Gouda, E.A.; Kotb, M.F.; Ghoneim, S.S.M.; Al-Harthi, M.M.; El-Fergany, A.A. Performance Assessment of Solar Generating Units Based on Coot Bird Metaheuristic Optimizer. IEEE Access 2021, 9, 111616–111632. [Google Scholar] [CrossRef]
  49. Mahdy, A.; Hasanien, H.-M.; Helmy, W.; Turky, R.-A.; Aleem, S.-H.-A. Transient Stability Improvement of Wave Energy Conversion Systems Connected to Power Grid Using Anti-Windup-Coot Optimization Strategy. Energy 2022, 245, 123321. [Google Scholar] [CrossRef]
  50. Houssein, E.-H.; Hashim, F.-A.; Ferahtia, S.; Rezk, H. Battery Parameter Identification Strategy Based on Modified Coot Optimization Algorithm. J. Energy Storage 2022, 46, 103848. [Google Scholar] [CrossRef]
  51. Alqahtani, A.-S.; Saravanan, P.; Maheswari, M.; Alshmrany, S. An Automatic Query Expansion Based on Hybrid CMO-COOT Algorithm for Optimized Information Retrieval. J. Supercomput. 2022, 78, 8625–8643. [Google Scholar] [CrossRef]
  52. Shan, L.; Qiang, H.; Li, J.; Wang, Z.-Q. Chaotic Optimization Algorithm Based on Tent Map. Control Decis. 2005, 20, 179–182. [Google Scholar]
  53. Li, Y.; Han, M.; Guo, Q. Modified Whale Optimization Algorithm Based on Tent Chaotic Mapping and Its Application in Structural Optimization. KSCE J. Civ. Eng. 2020, 24, 3703–3713. [Google Scholar] [CrossRef]
  54. Zhu, H.; Pu, B.; Zhu, Z.; Zhao, Y.; Song, Y. Two-dimensional Sine-tent-based Hyper Chaotic Map and Its Application in Image Encryption. J. Chin. Comput. Syst. 2019, 40, 1510–1518. [Google Scholar]
  55. Chechkin, A.V.; Metzler, R.; Klafter, J.; Gonchar, V.Y. Introduction to the Theory of Lévy Flights. In Anomalous Transport; Klages, R., Radons, G., Sokolov, I.M., Eds.; Wiley-VCH: Weinheim, Germany, 2008; pp. 129–162. [Google Scholar]
  56. Wang, X.-W.; Yan, Y.-X.; Gu, X.-S. Welding Robot Path Planning Based on Levy-PSO. Control Decis. 2017, 32, 373–377. [Google Scholar]
  57. Yang, X.-S.; Deb, S. Cuckoo Search via Lévy Flights. In Proceedings of the 2009 World Congress on Nature Biologically Inspired Computing (NaBIC), Coimbatore, India, 9–11 December 2009; pp. 210–214. [Google Scholar]
  58. Mantegna, R.N. Fast, Accurate Algorithm for Numerical Simulation of Lévy Stable Stochastic Processes. Phys. Rev. E 1994, 49, 4677–4683. [Google Scholar] [CrossRef]
  59. Haklı, H.; Uğuz, H. A Novel Particle Swarm Optimization Algorithm with Levy Flight. Appl. Soft Comput. 2014, 23, 333–345. [Google Scholar] [CrossRef]
  60. Tizhoosh, H.R. Opposition-Based Learning: A New Scheme for Machine Intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC’06), Vienna, Austria, 28–30 November 2005; Volume 1, pp. 695–701. [Google Scholar]
  61. He, Q.; Lin, J.; Xu, H. Hybrid Cauchy Mutation and Uniform Distribution of Grasshopper Optimization Algorithm. Control Decis. 2021, 36, 1558–1568. [Google Scholar]
  62. Wang, H.; Li, H.; Liu, Y.; Li, C.; Zeng, S. Opposition-Based Particle Swarm Algorithm with Cauchy Mutation. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; pp. 4750–4756. [Google Scholar]
  63. Higashi, N.; Iba, H. Particle Swarm Optimization with Gaussian Mutation. In Proceedings of the 2003 IEEE Swarm Intelligence Symposium, Indianapolis, IN, USA, 26 April 2003; pp. 72–79. [Google Scholar]
Figure 1. Chaotic tent map: (a) chaotic tent map bifurcation diagram; (b) Lyapunov exponential curve.
Figure 1. Chaotic tent map: (a) chaotic tent map bifurcation diagram; (b) Lyapunov exponential curve.
Sensors 22 03383 g001
Figure 2. Lévy flight simulation diagram.
Figure 2. Lévy flight simulation diagram.
Sensors 22 03383 g002
Figure 3. Relationship between arbitrary real numbers and their opposition-based learning numbers.
Figure 3. Relationship between arbitrary real numbers and their opposition-based learning numbers.
Sensors 22 03383 g003
Figure 4. Probability density function curves for the standard Cauchy and Gaussian distributions.
Figure 4. Probability density function curves for the standard Cauchy and Gaussian distributions.
Sensors 22 03383 g004
Figure 5. Flowchart of the COOTCLCO coverage optimization algorithm.
Figure 5. Flowchart of the COOTCLCO coverage optimization algorithm.
Sensors 22 03383 g005
Figure 6. Test function F3 and comparison of convergence curves on F3.
Figure 6. Test function F3 and comparison of convergence curves on F3.
Sensors 22 03383 g006
Figure 7. Test function F4 and comparison of convergence curves on F4.
Figure 7. Test function F4 and comparison of convergence curves on F4.
Sensors 22 03383 g007
Figure 8. Test function F8 and comparison of convergence curves on F8.
Figure 8. Test function F8 and comparison of convergence curves on F8.
Sensors 22 03383 g008
Figure 9. Test function F11 and comparison of convergence curves on F11.
Figure 9. Test function F11 and comparison of convergence curves on F11.
Sensors 22 03383 g009
Figure 10. Test function F12 and comparison of convergence curves on F12.
Figure 10. Test function F12 and comparison of convergence curves on F12.
Sensors 22 03383 g010
Figure 11. Test function F20 and comparison of convergence curves on F20.
Figure 11. Test function F20 and comparison of convergence curves on F20.
Sensors 22 03383 g011
Figure 12. Test function F21 and comparison of convergence curves on F21.
Figure 12. Test function F21 and comparison of convergence curves on F21.
Sensors 22 03383 g012
Figure 13. Comparison of the average coverage rate curves of the seven algorithms in the three cases: (a) q = 25, Rs = 10 m, Iteration = 1500; (b) q = 35, Rs = 10 m, Iteration = 1500; (c) q = 45, Rs = 10 m, Iteration = 1500.
Figure 13. Comparison of the average coverage rate curves of the seven algorithms in the three cases: (a) q = 25, Rs = 10 m, Iteration = 1500; (b) q = 35, Rs = 10 m, Iteration = 1500; (c) q = 45, Rs = 10 m, Iteration = 1500.
Sensors 22 03383 g013
Figure 14. Comparison of average coverage rates under different numbers of nodes.
Figure 14. Comparison of average coverage rates under different numbers of nodes.
Sensors 22 03383 g014
Figure 15. Initial coverage diagram and coverage optimization diagram of COOTCLCO in three cases: (a) q = 25, Rs = 10 m, Iteration = 1500; (b) q = 35, Rs = 10 m, Iteration = 1500; (c) q = 45, Rs = 10 m, Iteration = 1500.
Figure 15. Initial coverage diagram and coverage optimization diagram of COOTCLCO in three cases: (a) q = 25, Rs = 10 m, Iteration = 1500; (b) q = 35, Rs = 10 m, Iteration = 1500; (c) q = 45, Rs = 10 m, Iteration = 1500.
Sensors 22 03383 g015aSensors 22 03383 g015b
Figure 16. Comparison of coverage rates before and after optimization by COOTCLCO under different numbers of nodes.
Figure 16. Comparison of coverage rates before and after optimization by COOTCLCO under different numbers of nodes.
Sensors 22 03383 g016
Figure 17. Initial coverage diagram and coverage optimization diagram of COOTCLCO in three cases: (a) q = 45, Rs = 10 m, Iteration = 500; (b) q = 45, Rs = 10 m, Iteration = 1000; (c) q = 45, Rs = 10 m, Iteration = 1500.
Figure 17. Initial coverage diagram and coverage optimization diagram of COOTCLCO in three cases: (a) q = 45, Rs = 10 m, Iteration = 500; (b) q = 45, Rs = 10 m, Iteration = 1000; (c) q = 45, Rs = 10 m, Iteration = 1500.
Sensors 22 03383 g017aSensors 22 03383 g017b
Figure 18. Comparison of coverage rates before and after optimization by COOTCLCO when the number of nodes is 45.
Figure 18. Comparison of coverage rates before and after optimization by COOTCLCO when the number of nodes is 45.
Sensors 22 03383 g018
Table 1. Notation descriptions.
Table 1. Notation descriptions.
NotationsDescriptions
qNumber of sensor nodes
nNumber of target monitoring points
M × NSize of the monitoring area
SiThe i-th sensor node
xi, yiThe location coordinates of each sensor node
TjThe j-th target monitoring point
xj, yjThe location coordinates of each target point
RsSensing radius
RcCommunication radius
d(Si, Tj)The Euclidean distance between Si and Tj
pProbability of monitoring points being covered by nodes
PProbability of monitoring points being jointly sensed
CovCoverage rate
CootPos(i)The position of the i-th COOT
LeaderPos(k)The position of the selected leader
dThe number of variables or problem dimensions
ubThe upper bound of the search space
lbThe lower bound of the search space
QRandom initialization of the location
A, BControl parameters
NLThe number of leaders
R1, R2, R3, R4The random numbers between the interval [0, 1]
RThe random number between the interval [−1, 1]
α, µ, γ, λ, vControl parameters
sSearch path of the Lévy flight
XLeaderPos(i)The inverse solution of the current leader position
max_IterMaximum number of iterations
XGbestNewThe latest position after perturbed by Cauchy mutation
PsThe selection probability
Table 2. Unimodal benchmark functions.
Table 2. Unimodal benchmark functions.
FFunctionDimRangefmin
F1 f 1 ( x ) = i = 1 n x i 2 30[−100, 100]0
F2 f 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | 30[−10, 10]0
F3 f 3 ( x ) = i = 1 n ( j 1 i x j ) 2 30[−100, 100]0
F4 f 4 ( x ) = max { | x i | , 1 i n } 30[−100, 100]0
F5 f 5 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] 30[−30, 30]0
F6 f 6 ( x ) = i = 1 n ( [ x i + 0.5 ] ) 2 30[−100, 100]0
F7 f 7 ( x ) = max { | x i | , 1 i n } 30[−1.28, 1.28]0
Table 3. Multimodal benchmark functions.
Table 3. Multimodal benchmark functions.
FFunctionDimRangefmin
F8 f 8 ( x ) = i = 1 n x i sin ( | x i | ) 30[−500, 500]−418.9829×n
F9 f 9 ( x ) = i = 1 n x i 2 10 cos ( 2 π x i ) + 10 30[−5.12, 5.12]0
F10 f 10 ( x ) = 20 exp ( 0.2 1 n i = 1 n x i 2 ) exp ( 1 n i n cos ( 2 π x i ) ) + 20 + e 30[−32, 32]0
F11 f 11 ( x ) = i = 1 n x i 2 4000 i = 1 n cos ( x i i ) 1 30[−600, 600]0
F12 f 12 ( x ) = π n { 10 sin 2 ( π y 1 ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + 10 sin 2 ( π y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n u ( x i , 10 , 100 , 4 ) y i = 1 + x i + 1 4 ,   u ( x i , a , k , m ) = { k ( x i a ) m x i > a 0 a < x i < a k ( x i a ) m x i < a } 30[−50, 50]0
F13 f 13 ( x ) = 0.1 { sin 2 ( 3 π x i ) + i = 1 n ( x i 1 ) 2 [ 1 + sin 2 ( 3 π x i + 1 ) ] + ( x n 1 ) 2 [ 1 + sin 2 ( 2 π x n ) ] } + i = 1 n u ( x i , 5 , 100 , 4 ) 30[−50, 50]0
Table 4. Fixed-dimension multimodal benchmark functions.
Table 4. Fixed-dimension multimodal benchmark functions.
FFunctionDimRangefmin
F14 f 14 ( x ) = ( 1 500 + j = 1 25 1 j + i = 1 2 ( x i a i j ) 6 ) 1 2[−65, 65]1
F15 f 15 ( x ) = i = 1 11 [ a i x 1 ( b i 2 + b i x 2 ) b i 2 + b i x 3 + x 4 ] 2 4[−5, 5]0.0003
F16 f 16 ( x ) = 4 x 1 2 2.1 x 1 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 2[−5, 5]−1.0316
F17 f 17 ( x ) = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) cos x 1 + 10 2[−5, 5]0.398
F18 f 18 ( x ) = [ 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ] × [ 30 + ( 2 x 1 3 x 2 ) 2 × ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ] 2[−2, 2]3
F19 f 19 ( x ) = i = 1 4 c i exp ( j = 1 3 a i j ( x j p i j ) 2 ) 3[1, 3]−3.86
F20 f 20 ( x ) = i = 1 4 c i exp ( j = 1 6 a i j ( x j p i j ) 2 ) 6[0, 1]−3.32
F21 f 21 ( x ) = i = 1 5 [ ( X a i ) ( X a i ) T + c i ] 1 4[0, 10]−10.1532
F22 f 22 ( x ) = i = 1 7 [ ( X a i ) ( X a i ) T + c i ] 1 4[0, 10]−10.4028
F23 f 23 ( x ) = i = 1 10 [ ( X a i ) ( X a i ) T + c i ] 1 4[0, 10]−10.5363
Table 5. Parameter settings of the algorithm.
Table 5. Parameter settings of the algorithm.
AlgorithmsParametersValues
COOTCLCOPopulation30
Iteration1500
R[−1, 1]
R1[0, 1]
R2[0, 1]
µ2
rtan((rand( )−0.5) × 0.5)
COOTPopulation30
Iteration1500
R[−1, 1]
R1[0, 1]
R2[0, 1]
PSOPopulation30
Iteration1500
c1, c22
wmin0.2
wmax0.9
GWOPopulation30
Iteration1500
a[2, 0]
SSAPopulation30
Iteration1500
c1, c2, c3[0, 1]
BOAPopulation30
Iteration1500
a0.1
c0.01
p0.6
SOAPopulation30
Iteration1500
A[2, 0]
fc2
SCAPopulation30
Iteration1500
a2
r1, r2, r3, r4[0, 1]
Table 6. Results of unimodal and multimodal benchmark functions.
Table 6. Results of unimodal and multimodal benchmark functions.
FunctionCriteriaCOOTCLCOCOOTPSOGWOSSABOASOASCA
F1avg2.659 × 10−831.0898 × 10−314.6753 × 10−132.3081 × 10−908.514 × 10−092.4291 × 10−156.4176 × 10−430.00021788
std1.4561 × 10−825.9691 × 10−311.5343 × 10−129.858 × 10−901.546 × 10−091.6651 × 10−161.831 × 10−420.0010654
W/++++++
R24617538
F2avg6.2512 × 10−366.4608 × 10−211.2222 × 10−051.4253 × 10−520.810981.6096 × 10−123.8153 × 10−272.9312 × 10−08
std3.3024 × 10−353.5387 × 10−203.694 × 10−052.1289 × 10−521.07531.3002 × 10−139.4533 × 10−276.6456 × 10−08
W/+++++
R24718536
F3avg3.5233 × 10−776.9729 × 10−380.198343.1175 × 10−2330.52112.09 × 10−156.7979 × 10−232114.2643
std1.9298 × 10−763.8192 × 10−370.103061.573 × 10−2225.22641.5103 × 10−161.5973 × 10−222264.7997
W/+++++++
R12647538
F4avg1.7579 × 10−295.3006 × 10−220.0653093.1451 × 10−225.05171.8376 × 10−121.4693 × 10−1310.2213
std9.6209 × 10−292.8788 × 10−210.0268984.7313 × 10−222.72231.1907 × 10−133.593 × 10−138.0463
W/+++++
R12637548
F5avg27.803249.088330.602526.5966160.31228.904127.917340.7895
std0.2284864.958317.60350.91726312.65240.0257560.7336250.9841
W/+++++
R27518436
F6avg0.00270520.00114748.5311 × 10−130.644759.294 × 10−095.06553.21554.3367
std0.0016890.00069673.7915 × 10−120.345482.227 × 10−090.634970.440210.41554
W/++++
R43152867
F7avg0.00122010.00160250.0297910.000521730.062150.000636140.000716360.018585
std0.00109150.00127230.00956640.000318040.0260990.000215720.000560610.016281
W/+++
R45718236
F8avg−9259.1292−7727.7499−2947.3405−6119.6079−7766.648−4520.5507−5435.957−3978.2828
std746.799843.0888528.7386453.4792689.3375324.7176662.0541259.1258
W/+++++++
R13842657
F9avg5.6843 × 10−154.3428 × 10−1247.85750.3853757.939729.87872.013811.9765
std1.7345 × 10−142.3765 × 10−1114.96511.466916.695268.32737.731421.3326
W/++++++
R12738645
F10avg4.1034 × 10−141.2819 × 10−139.0532 × 10−081.1191 × 10−141.86225.2254 × 10−1319.959316.2366
std1.5409 × 10−133.9915 × 10−131.7522 × 10−073.2788 × 10−150.876683.8682 × 10−130.00127397.4322
W/++++
R14526387
F11avg3.7007 × 10−174.9516 × 10−159.48590.00189450.0105813.7007 × 10−180.0022860.26667
std8.9073 × 10−172.6891 × 10−143.94010.00514350.0106282.027 × 10−170.00923260.27876
W/++++++
R23846157
F12avg2.5932 × 10−050.0701010.345820.0389755.42610.403240.28961.1145
std1.7952 × 10−050.217640.505760.0204644.51210.140440.13261.0823
W/+++++++
R13528647
F13avg0.00851880.0157340.000366250.450240.59242.47691.969824.3547
std0.0111350.0233160.0020060.24013.20580.389480.15777104.6906
W/++++++
R23145768
+/≈/−/8/4/111/2/07/3/312/0/110/1/210/2/113/0/0
Table 7. Results of fixed-dimension multimodal benchmark functions.
Table 7. Results of fixed-dimension multimodal benchmark functions.
FunctionCriteriaCOOTCLCOCOOTPSOGWOSSABOASOASCA
F14avg0.9980.9981.69064.19220.9981.06431.39481.3287
std3.0018×10−162.2395 × 10−161.49114.41831.725 × 10−160.362620.807210.75206
W/+++++
R12783465
F15avg0.000469610.000648390.00049610.0050880.000824780.000326090.00110460.00081571
std0.000207120.000318950.000396010.00857540.000244941.7582×10−050.000317810.00030811
W/+++++
R24386175
F16avg−1.0316−1.0316−1.0316−1.0316−1.0316−1.0316−1.0316−1.0316
std6.8377×10−161.8251 × 10−126.7752 × 10−162.3368 × 10−093.931 × 10−151.0339 × 10−051.4916 × 10−071.4874 × 10−05
W/+++
R14253867
F17avg0.397890.397890.397890.397890.397890.398230.39790.39841
std01.2974 × 10−1507.2275 × 10−082.070 × 10−150.000791941.0205 × 10−050.00052896
W/+++++
R14153768
F18avg333333.011233
std3.2769×10−151.5417 × 10−141.7916 × 10−157.3174 × 10−063.959 × 10−140.0321311.9517 × 10−054.1425 × 10−05
W/+++++
R14253876
F19avg−0.30048−0.30048−3.8628−0.30048−0.30048−0.30048−0.30048−0.30048
std2.2584 × 10−162.2584 × 10−162.7101×10−152.2584 × 10−162.259 × 10−162.2584 × 10−162.2584 × 10−162.2584 × 10−16
W/++++
R23174658
F20avg−3.2982−3.2943−3.2625−3.277−3.2263−3.1381−2.7975−2.8919
std0.048370.0511460.0604630.0735440.0486820.147290.546440.39914
W/++++++
R12435687
F21avg−9.6924−9.2305−6.3881−9.1395−7.5573−9.0466−3.1981−3.0402
std1.34212.13653.46662.06173.1220.949493.90742.2051
W/+++++++
R12635478
F22avg−9.5169−10.2271−6.578−10.4027−9.2863−9.4478−6.1925−3.9181
std2.01470.962923.51470.000137332.5841.13994.68271.9622
W/++++
R32615478
F23avg−9.7655−10.3577−8.1972−10.5362−9.4872−10.0631−8.1032−4.9455
std1.86980.978743.64660.000106172.43320.34343.90971.9771
W/+++
R42615378
+/≈/−/4/4/25/4/17/1/24/6/07/1/210/0/010/0/0
Table 8. Overall Wilcoxon signed-rank test results, average value rank results, and algorithm rank results.
Table 8. Overall Wilcoxon signed-rank test results, average value rank results, and algorithm rank results.
ResultCOOTCLCOCOOTPSOGWOSSABOASOASCA
+/≈/−/12/8/316/6/114/4/516/6/117/2/420/2/123/0/0
Average rank1.7833.0874.6963.5225.3914.9575.3486.957
Overall rank12437568
Table 9. Experimental parameter settings for the node deployment area.
Table 9. Experimental parameter settings for the node deployment area.
ParametersValues
Area of deployment100 m × 100 m
Sensing radius (Rs)10 m
Communication radius (Rc)20 m
Number of sensor nodes (q)25, 35, 45
Number of iterations (Iteration)500, 1000, 1500
Table 10. Comparison of average coverage rate.
Table 10. Comparison of average coverage rate.
Algorithmq = 25q = 35q = 45
Average Coverage Rate/%Average Coverage Rate/%Average Coverage Rate/%
COOTCLCO75.32990.33296.990
PSO72.99980.38387.336
BOA62.71874.63483.102
SOA69.06682.75290.802
WOA69.91383.94791.600
HHO73.83189.11595.680
BES73.45988.64394.978
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, Y.; Zhang, J.; Wei, W.; Qin, T.; Fan, Y.; Luo, X.; Yang, J. Research on Coverage Optimization in a WSN Based on an Improved COOT Bird Algorithm. Sensors 2022, 22, 3383. https://doi.org/10.3390/s22093383

AMA Style

Huang Y, Zhang J, Wei W, Qin T, Fan Y, Luo X, Yang J. Research on Coverage Optimization in a WSN Based on an Improved COOT Bird Algorithm. Sensors. 2022; 22(9):3383. https://doi.org/10.3390/s22093383

Chicago/Turabian Style

Huang, Yihui, Jing Zhang, Wei Wei, Tao Qin, Yuancheng Fan, Xuemei Luo, and Jing Yang. 2022. "Research on Coverage Optimization in a WSN Based on an Improved COOT Bird Algorithm" Sensors 22, no. 9: 3383. https://doi.org/10.3390/s22093383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop