Next Article in Journal
Complexity and Evolution
Next Article in Special Issue
A Hybrid Particle Swarm Optimization Algorithm with Dynamic Adjustment of Inertia Weight Based on a New Feature Selection Method to Optimize SVM Parameters
Previous Article in Journal
Efficient Video Watermarking Algorithm Based on Convolutional Neural Networks with Entropy-Based Information Mapper
Previous Article in Special Issue
Optimal Security Protection Strategy Selection Model Based on Q-Learning Particle Swarm Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diversity-Aware Marine Predators Algorithm for Task Scheduling in Cloud Computing

School of Electronics and Information Engineering, Nanjing University of Information Science & Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2023, 25(2), 285; https://doi.org/10.3390/e25020285
Submission received: 29 November 2022 / Revised: 31 January 2023 / Accepted: 31 January 2023 / Published: 2 February 2023
(This article belongs to the Special Issue Information Theory and Swarm Optimization in Decision and Control)

Abstract

:
With the increase in cloud users and internet of things (IoT) applications, advanced task scheduling (TS) methods are required to reasonably schedule tasks in cloud computing. This study proposes a diversity-aware marine predators algorithm (DAMPA) for solving TS in cloud computing. In DAMPA, to enhance the premature convergence avoidance ability, the predator crowding degree ranking and comprehensive learning strategies were adopted in the second stage to maintain the population diversity and thereby inhibit premature convergence. Additionally, a stage-independent control of the stepsize-scaling strategy that uses different control parameters in three stages was designed to balance the exploration and exploitation abilities. Two case experiments were conducted to evaluate the proposed algorithm. Compared with the latest algorithm, in the first case, DAMPA reduced the makespan and energy consumption by 21.06% and 23.47% at most, respectively. In the second case, the makespan and energy consumption are reduced by 34.35% and 38.60% on average, respectively. Meanwhile, the algorithm achieved greater throughput in both cases.

1. Introduction

Cloud computing is a large resource pool that is dynamic and scalable, and the data center of a third-party service operator provides resources. Users can directly use the computing and storage resources of the cloud servers through the network [1], and internet of things (IoT) applications are deployed on the cloud. It has facilitated applications based on artificial intelligence and IoT [2], whereas, with the many conveniences brought by cloud computing, the increasing task to be processed and expansion of cloud resources also make scheduling cloud resources challenging.
Cloud resource scheduling has two main layers. The first layer schedules appropriate virtual resources for tasks submitted by users, and the second layer schedules appropriate hosts for virtual resources. This study focuses on the first layer because task scheduling in cloud computing (TSCC) directly affects the quality of services (QoS) parameters [3], such as the makespan, energy consumption, resource utilization rate, task response time, and task rejection rate. Under the constraints of QoS parameters, mapping a set of tasks to a suitable virtual resource is an NP-hard problem [4,5], and the algorithm suffers from dimensionality breakdown as the problem sizes increase.
Hence, ideas that apply meta-heuristic algorithms (MHAs) to TS have emerged to efficiently allocate available resources to complex and diverse incoming tasks within a reasonable time and with limited resources because of some of its inherent properties, such as its stochastic behavior. In addition, it has no dependency on the problem being solved and can search the solution space quickly to find the approximate optimal solution. Classical meta-heuristics, such as GA [6], PSO [7], novel whale optimization algorithm (WOA) [8], and Harris hawks optimization (HHO) [9] have been applied to TSCC successfully.
Although many MHA have been applied to cloud computing task scheduling problems, these algorithms tend to fall into the local optimum, resulting in high energy consumption of the system, long task completion time and other problems, affecting the overall optimization effect. This study applies a novel MHA, the marine predators algorithm (MPA), to solve TS in cloud computing. The motivation for this is that the MPA algorithm is capable of avoiding falling into the local optimum and achieves excellent performance in the optimization of complex problems [10]. To further strengthen the performance of the MPA algorithm and find a better solution that satisfies the QoS parameters, we propose a diversity-aware marine predators algorithm (DAMPA) to reduce makespan and energy consumption and increase throughput. Two strategies were used in the DAMPA to raise the variety of the population and prevent premature convergence. First, the predator crowding degree ranking strategy was designed to determine whether a predator performs exploration or exploitation in the second stage. Second, a comprehensive learning strategy was applied to enable predators to share the best experience. In addition, a stage-independent control of the stepsize-scaling strategy in DAMPA, which uses different control parameters in three stages, was designed to balance the exploration and exploitation abilities. The main contributions of this study are described below:
(1)
The DAMPA is proposed for resolving TSCC in reducing makespan and energy consumption and increasing throughput.
(2)
To avoid premature convergence, the predator crowding degree ranking strategy is designed, and the comprehensive learning strategy is applied.
(3)
A stage-independent control of the stepsize-scaling strategy is designed to balance the exploration and exploitation abilities.
This article’s structure is as follows: Section 2 discusses the similar works on TSCC. Section 3 introduces the problem formulation of the TS. Section 4 first introduces the MPA algorithm and then elaborates on our proposed DAMPA algorithm. Section 5 presents an experimental evaluation consisting of tests for multiple indicators. Section 6 is the conclusion of this work.

2. Related Works

TSCC has received considerable attention from scholars. Marahatta et al. [11] proposed a scheduling approach that first classifies heterogeneous tasks and virtual machines (VMs), then similar types of tasks are combined and scheduled. Additionally, it exploits the energy efficiency and optimal operating frequency of heterogeneous physical hosts to save energy when creating and deleting VMs. Hussain et al. [12] divided TS into two stages according to the optimization objectives. The first stage of scheduling focuses on reducing the execution time of the tasks, and the second stage is to reduce energy consumption.
In addition to classification-based and phased scheduling methods, meta-heuristic algorithms are key in solving TS problems in cloud computing because of their superior performance in the optimization of complex problems. For example, Chen et al. [13] used the WOA algorithm for improving the efficiency of task execution. Abdullah et al. [14] combined the PSO algorithm with the Pareto optimal frontier, and a mutation operator was introduced to avoid premature convergence. Laili et al. [15] designed a parallel TS algorithm, which divides tasks into several groups, finds the best solution with meta-heuristic algorithms for each group and then merges all sub-solutions into the final solution. This algorithm decreases the total execution time and energy consumption, but the TS model is complex. Ali et al. [16] proposed an optimization model based on NSGA-II that realizes automatic mapping between tasks and cloud nodes and minimizes the total execution time.
To reduce the makespan, Xiong et al. [17] introduced Johnson’s rule to GA, which adds a new crossover and the mutation operation. Pirozmand et al. [18] first prioritized tasks and then assigned tasks to VMs using GA. Pang et al. [19] introduced the estimated distribution algorithm (EDA) to the GA algorithm to initialize the solution. Xu et al. [20] prioritized the TS sequence according to the priority algorithm and used the ant colony algorithm (ACO) to obtain a global optimal scheduling scheme under certain constraints.
Because of the limitations of a single meta-heuristic algorithm, several scholars have adopted a hybrid bio-inspired algorithm. For example, Attiya et al. [21] combined the manta-ray foraging optimizer (MRFO) and Salp Swarm algorithm (SSA), which strengthens the exploitation capability, and the results indicate that the approach outperforms the existing algorithms. Similarly, Walia et al. [22] combined GA with the flower pollination algorithm (FPA), and the crossover and mutation of the GA are introduced in the FPA. Domanal et al. [23] used an improved PSO to find a better solution for mapping tasks to VMs and adopted a modified hybrid bio-inspired algorithm for resource allocation according to the needs of the tasks. Fu et al. [24] combined the PSO and GA algorithms to reduce the task completion time. Although hybrid algorithms have achieved better results in several indicators, they are also overly complex.
Based on the above analysis, various algorithms have been applied to the TS. The optimization of makespan, energy consumption, and throughput, three important indicators of TS, is still insufficient. Therefore, a new method based on MPA was proposed for TSCC.

3. Problem Formulation

3.1. Task Scheduling

In the cloud, the data center receives the tasks uploaded through the network, and the broker allocates the tasks to the appropriate available resources according to the task requirements and information of the VMs and optimization goals. TS in the cloud can be summarized as the mapping of a group of tasks T 1 , T 2 , T 3 , T m to V M 1 , V M 2 , V M 3 , V M n , m is the total number of tasks to be processed, n is the total number of virtual machines. T i has a corresponding task size, and the key factors of V M j are processing speed, RAM, and CPU processing elements. One task can only be assigned to one VM, and each VM executes multiple tasks. In this study, our goal is to map tasks to VMs to decrease the makespan and energy consumption and increase throughput.

3.2. Mathematical Model

In the TS process, D i j represents the decision variable of the allocation process, which is expressed as follows:
D i j = 1 , T i to V M j 0 , otherwise
The ranges of the subscripts i and j of the decision variable D i j are the number of tasks and the number of virtual machines respectively. When the number of virtual machines or tasks increases, more decision values are required. Since each task must be handled by just one VM, and since the schedule allocates all the tasks, we have
j = 1 n D i j = 1
i = 1 m j = 1 n D i j = m
The processing time of T i on V M j can be calculated by the following:
t i j = W i S j
where t i j represents the processing time of T i on V M j , W i represents the task size of task T i , S j represents the processing speed of V M j .

3.2.1. Makespan

We assumed that each VM starts executing the first task from time 0 until the last task is executed, and the time spent during this period is the total execution time of V M j . The makespan can be calculated as follows:
t j = i = 1 m t i j · D i j
t M = max ( t j ) , j = 1 , 2 n .
where t j represents the total execution time of V M j , t M represents the makespan.

3.2.2. Energy Consumption

The energy consumed by VMs includes the idle state and the working state. The consumption of VMs in the idle state is approximately 0.6 times the working state [25]. The total energy consumption (TEC) is calculated as follows:
T E j = ( t j · α + ( t M t j ) · β ) · S j
T E C = j = 1 n T E j
where T E j is the energy consumption of the V M j , α = 10 8 · S j 2 , β = 0.6 · α .

3.2.3. Throughput

Throughput metric measures, the performance of the cloud system to complete tasks over a period is reflected in work completed by the cloud data center per unit of time. It can be calculated as follows:
T h r o u g h p u t = T a s k q u a n t i t y t M

3.2.4. Fitness Function

The fitness function measures the performance of individuals. Based on our optimization goals of energy consumption, makespan, and throughput, as makespan is a key factor in throughput, we set the objective function to the following to evaluate the candidate solutions:  
F i t n e s s = λ 1 · T E C + λ 2 · t M
where λ 1 + λ 2 = 1 , we set λ 1 = 0.5 and λ 2 = 0.5.

4. The Proposed Algorithm

This part first introduces the MPA algorithm to present its optimization process, and then the proposed DAMPA algorithm is elaborated. There are three differences between DAMPA and MPA: the predator crowding degree ranking strategy, comprehensive learning strategy, and stage-independent control of the stepsize-scaling strategy.

4.1. The Overview of MPA

4.1.1. Elite and Prey Matrix

Based on the theory of survival of the fittest, the best predators are better at foraging in natural survival. Therefore, the optimal solution is set as the top predator. The optimal solution for each predator constructs the E l matrix. The E l matrix is as follows:
E l = I 1 , 1 E I 1 , 2 E I 1 , d E I 2 , 1 E I 2 , 2 E I 2 , d E I k , 1 E I k , 2 E I k , d E
where I i E = I i , 1 E , I i , 2 E I i , d E , k marks the number of predators in the population, and dimension d represents the number of variables in the solution I i E .
The prey matrix ( P e ) and the E l matrix are similar, the P e matrix saves the new value produced each iteration, and the E l matrix represents the historical optimal solution of the predators. The P e matrix is expressed as follows:
P e = I 1 , 1 P I 1 , 2 P I 1 , d P I 2 , 1 P I 2 , 2 P I 2 , d P I k , 1 P I k , 2 P I k , d P
where I j P = I j , 1 P , I j , 2 P I j , d P . In the initialization phase of the algorithm, the values in the P e matrix will be randomly generated within the specified range, the optimal solution I E will be obtained through calculating each I P , and the E l matrix will be initialized by duplicating I E .

4.1.2. Optimization Process

In the first stage, all predators perform exploration. Taking advantage of the Brownian motion step sizes helps predators approach the optimal target more quickly at the initial stage when they are far from the optimal target. Formulated as follows, when i t e r < 1 3 M a x i t e r ,
L i = M B ( E l i M B P e i )
P e i = P e i + P R L i
where L i represents the step size of each predator, M B is a vector composed of standard normal distribution random numbers (RND), R is a uniform RND vector in [ 0 , 1 ] , and P is a constant, P e i and E l i are the row vector of the matrix, i t e r is the current number of iterations, and M a x i t e r is the maximum number of iterations.
The second stage involves the shift from exploration to exploitation. Half of the predators adopted the exploitation strategy, and the remainder adopted the exploration strategy. When 1 3 M a x i t e r < i t e r < 2 3 M a x i t e r , we have the following.
First one-half of predators:
L i = M L ( E l i M L P e i )
P e i = P e i + P R L i
Other predators:
L i = M B ( M B E l i P e i )
P e i = E l i + P C F L i
where M L is a RND vector generated based on the Levy distribution and C F controls the step sizes of the predators and can be calculated as follows:
C F = ( 1 i t e r M a x i t e r ) 2 i t e r M a x i t e r
In the third stage, all individuals are in a state of Levy movement and execute the exploitation strategy, when i t e r > 2 3 M a x i t e r :
L i = M L ( M L E l i P e i )
P e i = E l i + P C F L i
In simulating the survival of predators, the eddy taking shape or fish aggregation device (FADs) effect in the environment also causes predators to adjust to their movement states. A broad range of jumps helps the algorithm avoid stagnancy in the local optima. The FADs effect is expressed as follows:
P e i = P e i + C F P e l + R ( P e u P e l ) U , if r F A D s P e i + F A D s ( 1 r ) + r ( P e i 1 P e i 2 ) , if r > F A D s
where F A D s = 0.2 represents the probability that the solution is affected during the update process, r is a RND in [0,1], and R is a RND vector in [0,1]. P e l consists of the minimum value in each dimension and P e u consists of the maximum value in each dimension. U is a binary number, and i 1 and i 2 are random indices. Finally, the pseudocode of the MPA is expressed as Algorithm 1.
Algorithm 1 MPA
1:
initialize the Pe, S , P, FADs, Maxiter, Tf, Tp
2:
while  i t e r < M a x i t e r   do
3:
   Compute the fitness
4:
   Construct Elite matrix
5:
   Save memory
6:
   update CF using (19)
7:
   if  i t e r < 1 3 M a x i t e r  then
8:
      Update Pe using (13) and (14)
9:
   else if  1 3 M a x i t e r < i t e r < 2 3 M a x i t e r  then
10:
     First one-half of predators:
11:
     Update Pe using (15) and (16)
12:
     other predators:
13:
     Update Pe using (17) and (18)
14:
   else if  i t e r > 2 3 M a x i t e r  then
15:
     Update Pe using (20) and (21)
16:
   end if
17:
   Compute the fitness
18:
   Saving memory
19:
   Applying FADs effect using (22)
20:
   iter++
21:
end while

4.2. The DAMPA Algorithm

4.2.1. The Predator Crowding Degree Ranking Strategy

In the optimization process of the MPA, the second stage, which carries out the transition from exploration to exploitation, is indispensable. Here, the MPA algorithm neglects the population diversity when selecting predators to perform exploitation and exploration. If predators that are closer to each other are selected for exploitation in the second stage, the population diversity will be prematurely reduced. Therefore, to raise the variety of the population and thereby avoid premature convergence, the predator crowding degree ranking strategy is adopted in the DAMPA. The crowding degree of a predator describes the magnitude of the location difference between the predator and the remaining predators in the population, if a predator is close to other predators, it has a larger crowding degree. After the end of the first stage in the DAMPA, the crowding degree of each predator is calculated using (23) and (24), and rank the predators according to their degree of crowding from small to large. The predators with lower rankings execute the exploration, and the remaining predators execute the exploitation. The number of predators executing the two strategies was the same. Algorithm 2 shows the crowding degree ranking algorithm. We assume that P e i is the current predator and P e j is the other predator. The mathematical expression for crowding degree (CD) is as follows:
D E P e i = j = 1 m ( k = 1 d ( I i , k I j , k ) 2 ) 0.5 ( i j )
C D P e i = 1 P o p u l a t i o n S i z e 1 D E P e i
where P e i = I i , 1 P , I i , 2 P I i , d P , P e j = I j , 1 P , I j , 2 P I j , d P .
Algorithm 2 Crowding degree ranking algorithm.
1:
initialize the CD, Ranking (two arrays of length PopulationSize)
2:
for i = 0 to PopulationSize do
3:
   Computing C D i based on (23) and (24)
4:
end for
5:
for  j = 0 to PopulationSize do
6:
   for  k = 0 to PopulationSize do
7:
     if  C D j < C D k  then
8:
         R a n k i n g j + +
9:
     end if
10:
  end for
11:
end for
12:
return  R a n k i n g

4.2.2. Comprehensive Learning Strategy

In DAMPA, a comprehensive learning strategy is also applied in the second stage, which enables the exchange of best experiences between predators. This maintains the population diversity and avoids premature convergence [26]. Predators with high crowding degrees are given a larger learning rate to maximize the population diversity, and the learning rate of predators can be calculated by the following:
P e l i = c + e e x p ( 10 ( R a n k i n g i 1 ) P o p u l a t i o n S i z e 1 ) e x p ( 10 ) 1
where c = 0.05 , and e = 0.5 .
A comprehensive learning strategy is shown in Algorithm 3. First, the learning rate P e l of the predators was calculated. Subsequently, a random value r is generated, and predators P e k 1 and P e k 2 are randomly selected. When r < P e l , the fitness value of each predator was calculated. If the fitness value of P e k 1 is less than P e k 2 , then P e k 1 is used as an exemplar. Otherwise, P e k 2 was used as an example. If r > P e l , then the predator is used as an exemplar.
Next, in the second stage, we applied comprehensive learning, in which (15) and (17) are replaced by (26) and (28), respectively.
First one-half of predators:
L i = M L ( E l i M L P e i + c 1 R ( P e b F i P e i ) )
c 1 = 1 1 + e θ i t e r M a x i t e r + 2 ( i t e r M a x i t e r 1 ) 2
where P e b F i = e x e m p l a r , θ = 0.0001 .
Other predators:
L i = M B ( M B E l i P e i + c 2 R ( P e b F i P e i ) )
c 2 = 1 1 + e θ i t e r M a x i t e r + 2 ( i t e r M a x i t e r ) 2
Algorithm 3 Comprehensive learning strategy.
1:
for  i = 0 to PopulationSize do
2:
   Calculate P e l i using (25)
3:
   Select two random agents with indexs k1,k2
4:
   if  P l i > r  then
5:
     Calculate the fitness of P e k 1 and P e k 2
6:
     if  f i t n e s s P e k 1 < f i t n e s s P e k 2  then
7:
         e x e m p l a r = P e k 1
8:
     else
9:
         e x e m p l a r = P e k 2
10:
     end if
11:
   end if
12:
   if  P l i < r  then
13:
      e x e m p l a r = P e i
14:
   end if
15:
end for

4.2.3. Stage-Independent Control of Stepsize-Scaling Strategy

In MPA, P is the parameter of step size scaling, which controls the enlargement or reduction of the step sizes in three stages simultaneously. Figure 1 shows the convergence performance at each stage with different step sizes scaling control parameters. When P = 0.1 , the algorithm performs better in the first stage and has a strong exploration ability, but the exploitation ability is insufficient in the third stage. When P = 0.5 , the algorithm has a strong exploitation ability, but the exploration ability in the first stage is insufficient. Therefore, to balance the exploration and exploitation capabilities of the MPA algorithm, the stage-independent control of the stepsize-scaling strategy is designed in DAMPA to optimize the performance at each stage by using P 1 , P 2 , and P 3 to control the scaling of the step sizes of the three stages. The control parameter is set to P 1 in the first stage. The second stage corresponds to P 2 . In the third stage, we have P 3 .
Based on the above operations, we use (30), (31), (32), and (33) to replace (14), (16), (18), and (21), respectively. The complete DAMPA is presented in Algorithm 4.
P e i = P e i + P 1 R L i
P e i = P e i + P 2 R L i
P e i = E l i + P 2 C F L i
P e i = E l i + P 3 C F L i
Algorithm 4 DAMPA
1:
Initialize the Pe, S , P1, P2, P3, FADs, Maxiter, Tfs, Tp
2:
while  i t e r < M a x i t e r   do
3:
   Compute the fitness of P e
4:
   Construct El matrix
5:
   Save memory
6:
   update CF using (19)
7:
   if  i t e r < 1 3 M a x i t e r  then
8:
     Update Py using (13) and (30)
9:
   else if  i t e r = 1 3 M a x i t e r  then
10:
     Calculate the crowding degree ranking
11:
  else if  1 3 M a x i t e r < i t e r < 2 3 M a x i t e r  then
12:
     if  R a n k i n g P e i < 1 2 P o p u l a t i o n S i z e  then
13:
        Update Py using (26) and (31)
14:
     else
15:
        Update Py using (28) and (32)
16:
     end if
17:
  else if  i t e r > 2 3 M a x i t e r  then
18:
     Update Py using (20) and (33)
19:
  end if
20:
   Compute P e
21:
   Save memory
22:
   Apply FADs effect using (22)
23:
   iter++
24:
end while

4.2.4. Complexity Analysis

The complexity of the MPA algorithm is O ( T ( P d × P s + C × P s ) ) , the crowding degree ranking algorithm is O ( P s × P s × P d ) , and the comprehensive learning strategy is O ( 1 3 × T × P s ) . The DAMPA complexity is O ( T ( P d × P s + C × P s ) + P s × P s × P d ) , where T is the maximum iterations, P s is the amount of predators, C is the evaluated cost, and P d is the dimension of predators.

5. Experiment and Analysis

This section introduces the dataset settings, parameter settings, and experimental results compared with those of other algorithms. The proposed DAMPA is written based on the Java language, and the experimental computer specifications are inter-core i7-9700CPU@3.0GHZ, 32 GB RAM, Windows 10 64-bit operating system, Cloudsim4.0.

5.1. Data Set

The establishment of the dataset was mainly from the perspectives of tasks and VMs. It was considered in two cases to simulate resource-limited and resource-rich situations.

5.1.1. Case1

The number of VMs is fixed at 50, and the processing speed of each VM to 2000 + j × 40 , where j is the index of the VMs and j = 1 , 2 , 3 50 . The tasks is set to 100, 200, 300, 400, 500, 600, 700, and 800. The size of the tasks was randomly generated within [200, 12,000].

5.1.2. Case2

The number of tasks is fixed at 200, and the size of each task was set to 1000 + i × 5 , where i is the index of the tasks and i = 1 , 2 , 3 200 . The VM is set to 50, 60, 70, 80, 90, 100, 110, 120, and 130. The S j of the VMs was generated in [200, 12,000] randomly.

5.2. Parameter Setting

We compared the proposed algorithm with existing algorithms including IMMPA [27], WOA, MRFOSSA [21], and HHO to verify our algorithm performance. Table 1 lists the parameter settings of the algorithms, the population size is 50, and the maximum iteration is ten thousand. Each algorithm was run twenty times independently, and the results were averaged.

5.3. Discretization

At this stage, since the proposed optimization algorithm contains continuous values, it needs to be discretized. The continuous values of predator need to be converted to discrete values (VMs number). First, the predator vector is normalized as follows:
normalized N V = N V min max min
where min and max are the minimum and maximum values in the predator vector I , respectively. NV is the new value generated by each update in the vector I , and normalized N V represents the normalized value. After that, the following equation is used to scale the value in vector I :
scaled N V = normalized N V ( n 1 ) + 1
where scaled N V is the scaled value, n is the number of virtual machines.

5.4. Experimental Results

In the first case, the fitness under different numbers of tasks is shown in Table 2, which indicates that DAMPA obtains a lower fitness value than IMMPA, HHO, MRFOSSA, and WOA. DAMPA has a small improvement based on IMMPA, and a large improvement based on HHO, MRFOSSA, and WOA. Table 3 shows the makespan under the different number of tasks. The makespan value increases when the amount of tasks to be processed increases. DAMPA obtains the lesser makespan value. To more accurately describe the effect of optimization, Table 4 compares the percentage of makespan decrease of different algorithms, when the task size is minimal and the makespan value is reduced by 2.60%, 11.56%, 16.27%, and 31.37%, over IMMPA, HHO, MRFOSSA, and WOA. With the expansion of the scale of the scheduling problem, finding a solution becomes difficult for all algorithms, and the overall optimization effect of DAMPA is constantly weakening. DAMPA improved by 7.38% at most based on IMMPA and improved by 1.86–11.56%, 12.59–21.06%, 7.58–31.37% based on HHO, MRFOSSA, and WOA respectively. Table 5 lists the TEC under the different numbers of tasks. More energy is consumed with an increasing number of tasks. DAMPA consumes less energy than IMMPA, HHO, MRFOSSA, and WOA. More specifically, Table 6 describes the improvement percentage in TEC. When the task size is minimal, the TEC is reduced by 0.99%, 17.64%, 23.47%, and 33.14% over IMMPA, HHO, MRFOSSA, and WOA, respectively. As the scale of the problem increases, the optimization effect of DAMPA weakens on the whole. Compared with IMMPA, HHO, MRFOSSA, and WOA, the improvement ranges are 0.07–6.33%, 1.85–17.64%, 11.57–23.47%, and 5.60–33.14%, respectively. The throughput of the system under different numbers of tasks is depicted in Figure 2, which shows that DAMPA achieves a greater throughput than IMMPA, HHO, MRFOSSA, and WOA.
In the second case, Figure 3 shows the effect of different numbers of VMs on fitness. The fitness value keeps increasing with the number of VMs increasing, which is caused by more energy consumption. The fitness value of DAMPA is below IMMPA, HHO, MRFOSSA, and WOA. The makespan for different numbers of VMs is shown in Figure 4, which indicates that DAMPA obtains a lower makespan value than IMMPA, HHO, MRFOSSA, and WOA. More VMs enable DAMPA, IMMPA, MRFOSSA, and HHO to find better scheduling schemes, makespan continues to decrease, and the WOA algorithm has the largest fluctuation. Table 7 lists the percentage reduction of makespan, the increase in VMs scale has no obvious impact on the optimization effect of DAMPA, the makespan was reduced by 6.81%, 26.92%, 34.35%, and 50.60% on average over IMMPA, HHO, MRFOSSA, and WOA, respectively. The TEC of VMs is shown in Figure 5, which shows that the DAMPA consumes less energy. To obtain more accurate results, Table 8 lists the percentage of TEC optimization, the TEC is reduced by 6.30%, 27.04%, 38.60%, and 47.50% on average over IMMPA, HHO, MRFOSSA, and WOA, respectively. Figure 6 shows the system throughput for different VMs. The DAMPA achieves a greater throughput of all algorithms.
Summarizing case 1 and case 2, when there are many tasks with limited resources, the improvement of DAMPA based on other algorithms is up to 33.14% at most, when the available resources increase, DAMPA can reduce the makespan by 50.60% and the energy consumption by 47.50% at most.

6. Conclusions

This study proposes a novel meta-heuristic algorithm DAMPA for solving TSCC. In DAMPA, the predator crowding degree ranking strategy and comprehensive learning strategy are taken to maintain the diversity of the population, thereby avoiding premature convergence. In the second stage of DAMPA, the predators are selected to perform exploration or exploitation according to the ranking of the crowding degree of the predators, and a comprehensive learning strategy makes predators share the optimal historical experience. Additionally, to balance the exploitation and exploration capabilities of the algorithm, a stage-independent control of the stepsize-scaling strategy is designed, which uses different control parameters for scaling the step sizes in three stages. Two case experimental results show that DAMPA achieves lower makespan and energy consumption and greater throughput compared with the latest algorithms: IMMPA, HHO, MRFOSSA, and WOA. Especially in the second case, DAMPA has a distinct advantage in solving TSCC problems.
In the future, we will focus on applying the DAMPA algorithm to other problems, such as cloud-fog collaborative TS with constraints and cloud-edge collaborative TS.

Author Contributions

Conceptualization, D.C. and Y.Z.; methodology, D.C.; software, D.C.; validation, D.C. and Y.Z.; writing—original draft preparation, D.C.; writing—review and editing, Y.Z.; visualization, D.C.; supervision, D.C.; project administration, Y.Z.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The research is supported by the National Key Research and Development Program of China (No. 2022YFB4401301) and National Natural Science Foundation of China (No. 61705109).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rimal, B.P.; Maier, M. Workflow Scheduling in Multi-Tenant Cloud Computing Environments. IEEE Trans. Parallel Distrib. Syst. 2017, 28, 290–304. [Google Scholar] [CrossRef]
  2. Abualigah, L.; Diabat, A.; Sumari, P.; Gandomi, A.H. Applications, Deployments, and Integration of Internet of Drones (IoD): A Review. IEEE Sens. J. 2021, 21, 25532–25546. [Google Scholar] [CrossRef]
  3. Houssein, E.H.; Gad, A.G.; Wazery, Y.M.; Suganthan, P.N. Task Scheduling in Cloud Computing Based on Meta-Heuristics: Review, Taxonomy, Open Challenges, and Future Trends. Swarm Evol. Comput. 2021, 62, 100841. [Google Scholar] [CrossRef]
  4. Yang, Y.; Shen, H. Deep Reinforcement Learning Enhanced Greedy Algorithm for Online Scheduling of Batched Tasks in Cloud in Cloud HPC Systems. IEEE Trans. Parallel Distrib. Syst. 2021, 33, 3003–3014. [Google Scholar] [CrossRef]
  5. Velliangiri, S.; Karthikeyan, P.; Arul Xavier, V.M.; Baswaraj, D. Hybrid Electro Search with Genetic Algorithm for Task Scheduling in Cloud Computing. Ain Shams Eng. J. 2021, 12, 631–639. [Google Scholar] [CrossRef]
  6. Shukla, D.K.; Kumar, D.; Kushwaha, D.S. WITHDRAWN: Task Scheduling to Reduce Energy Consumption and Makespan of Cloud Computing Using NSGA-II. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  7. Cui, Z.; Zhang, J.; Wu, D.; Cai, X.; Wang, H.; Zhang, W.; Chen, J. Hybrid Many-Objective Particle Swarm Optimization Algorithm for Green Coal Production Problem. Inf. Sci. 2020, 518, 256–271. [Google Scholar] [CrossRef]
  8. Abd Elaziz, M.; Attiya, I. An Improved Henry Gas Solubility Optimization Algorithm for Task Scheduling in Cloud Computing. Artif. Intell. Rev. 2021, 54, 3599–3637. [Google Scholar] [CrossRef]
  9. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris Hawks Optimization: Algorithm and Applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  10. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine Predators Algorithm: A Nature-Inspired Metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  11. Marahatta, A.; Pirbhulal, S.; Zhang, F.; Parizi, R.M.; Choo, K.-K.R.; Liu, Z. Classification-Based and Energy-Efficient Dynamic Task Scheduling Scheme for Virtualized Cloud Data Center. IEEE Trans. Cloud Comput. 2021, 9, 1376–1390. [Google Scholar] [CrossRef]
  12. Hussain, M.; Wei, L.-F.; Lakhan, A.; Wali, S.; Ali, S.; Hussain, A. Energy and Performance-Efficient Task Scheduling in Heterogeneous Virtualized Cloud Computing. Sustain. Comput. Inform. Syst. 2021, 30, 100517. [Google Scholar] [CrossRef]
  13. Chen, X.; Cheng, L.; Liu, C.; Liu, Q.; Liu, J.; Mao, Y.; Murphy, J. A WOA-Based Optimization Approach for Task Scheduling in Cloud Computing Systems. IEEE Syst. J. 2020, 14, 3117–3128. [Google Scholar] [CrossRef]
  14. Abdullah, M.; Al-Muta’a, E.A.; Al-Sanabani, M. Integrated MOPSO Algorithms for Task Scheduling in Cloud Computing. IFS 2019, 36, 1823–1836. [Google Scholar] [CrossRef]
  15. Laili, Y.; Guo, F.; Ren, L.; Li, X.; Li, Y.; Zhang, L. Parallel Scheduling of Large-Scale Tasks for Industrial Cloud-Edge Collaboration. IEEE Internet Things J. 2021, 3231–3242. [Google Scholar] [CrossRef]
  16. Ali, I.M.; Sallam, K.M.; Moustafa, N.; Chakraborty, R.; Ryan, M.J.; Choo, K.-K.R. An Automated Task Scheduling Model Using Non-Dominated Sorting Genetic Algorithm II for Fog-Cloud Systems. IEEE Trans. Cloud Comput. 2020, 10, 2294–2308. [Google Scholar] [CrossRef]
  17. Xiong, Y.; Huang, S.; Wu, M.; She, J.; Jiang, K. A Johnson’s-Rule-Based Genetic Algorithm for Two-Stage-Task Scheduling Problem in Data-Centers of Cloud Computing. IEEE Trans. Cloud Comput. 2019, 7, 597–610. [Google Scholar] [CrossRef]
  18. Pirozmand, P.; Hosseinabadi, A.A.R.; Farrokhzad, M.; Sadeghilalimi, M.; Mirkamali, S.; Slowik, A. Multi-Objective Hybrid Genetic Algorithm for Task Scheduling Problem in Cloud Computing. Neural Comput. Appl. 2021, 33, 13075–13088. [Google Scholar] [CrossRef]
  19. Pang, S.; Li, W.; He, H.; Shan, Z.; Wang, X. An EDA-GA Hybrid Algorithm for Multi-Objective Task Scheduling in Cloud Computing. IEEE Access 2019, 7, 146379–146389. [Google Scholar] [CrossRef]
  20. Xu, J.; Hao, Z.; Zhang, R.; Sun, X. A Method Based on the Combination of Laxity and Ant Colony System for Cloud-Fog Task Scheduling. IEEE Access 2019, 7, 116218–116226. [Google Scholar] [CrossRef]
  21. Attiya, I.; Elaziz, M.A.; Abualigah, L.; Nguyen, T.N.; El-Latif, A.A.A. An Improved Hybrid Swarm Intelligence for Scheduling IoT Application Tasks in the Cloud. IEEE Trans. Ind. Inf. 2022, 18, 6264–6272. [Google Scholar] [CrossRef]
  22. Walia, N.K.; Kaur, N.; Alowaidi, M.; Bhatia, K.S.; Mishra, S.; Sharma, N.K.; Sharma, S.K.; Kaur, H. An Energy-Efficient Hybrid Scheduling Algorithm for Task Scheduling in the Cloud Computing Environments. IEEE Access 2021, 9, 117325–117337. [Google Scholar] [CrossRef]
  23. Domanal, S.G.; Guddeti, R.M.R.; Buyya, R. A Hybrid Bio-Inspired Algorithm for Scheduling and Resource Management in Cloud Environment. IEEE Trans. Serv. Comput. 2020, 13, 3–15. [Google Scholar] [CrossRef]
  24. Fu, X.; Sun, Y.; Wang, H.; Li, H. Task Scheduling of Cloud Computing Based on Hybrid Particle Swarm Algorithm and Genetic Algorithm. Clust. Comput. 2021. [Google Scholar] [CrossRef]
  25. Abdel-Basset, M.; El-Shahat, D.; Elhoseny, M.; Song, H. Energy-Aware Metaheuristic Algorithm for Industrial-Internet-of-Things Task Scheduling Problems in Fog Computing Applications. IEEE Internet Things J. 2021, 8, 12638–12649. [Google Scholar] [CrossRef]
  26. Yousri, D.; Fathy, A.; Rezk, H. A New Comprehensive Learning Marine Predator Algorithm for Extracting the Optimal Parameters of Supercapacitor Model. J. Energy Storage 2021, 42, 103035. [Google Scholar] [CrossRef]
  27. Abdel-Basset, M.; Mohamed, R.; Elhoseny, M.; Bashir, A.K.; Jolfaei, A.; Kumar, N. Energy-Aware Marine Predators Algorithm for Task Scheduling in IoT-Based Fog Computing Applications. IEEE Trans. Ind. Inf. 2021, 17, 5068–5076. [Google Scholar] [CrossRef]
Figure 1. Phase convergence performance of different step sizes scaling control parameters.
Figure 1. Phase convergence performance of different step sizes scaling control parameters.
Entropy 25 00285 g001
Figure 2. Comparison based on throughput under different number of tasks.
Figure 2. Comparison based on throughput under different number of tasks.
Entropy 25 00285 g002
Figure 3. Comparison based on fitness under different numbers of VMs.
Figure 3. Comparison based on fitness under different numbers of VMs.
Entropy 25 00285 g003
Figure 4. Comparison based on makespan under different numbers of VMs.
Figure 4. Comparison based on makespan under different numbers of VMs.
Entropy 25 00285 g004
Figure 5. Comparison based on TEC under different numbers of VMs.
Figure 5. Comparison based on TEC under different numbers of VMs.
Entropy 25 00285 g005
Figure 6. Comparison based on throughput under different numbers of VMs.
Figure 6. Comparison based on throughput under different numbers of VMs.
Entropy 25 00285 g006
Table 1. Parameter setting.
Table 1. Parameter setting.
AlgorithmParameterValue
DAMPAP10.05
P20.5
P31
IMMPAP1
HHO β 1.5
WOAb1
MRFOSSAS2
Table 2. Comparison based on fitness under different number of tasks.
Table 2. Comparison based on fitness under different number of tasks.
TasksDAMPAIMMPAHHOMRFOSSAWOA
10011911233142615531800
20026032730285431673453
30037903898406545884491
40049895081526059395705
50062076251661472557167
60074907515794384968361
70085578665872596939093
8009710978110,03610,90810,676
Table 3. Comparison based on makespan under different number of tasks.
Table 3. Comparison based on makespan under different number of tasks.
TasksDAMPAIMMPAHHOMRFOSSAWOA
1008.168.379.229.7411.89
20015.1916.4016.8719.2420.82
30022.0023.1024.0826.7226.69
40028.5229.1429.8333.7032.55
50035.2135.7937.4440.2839.73
60041.7741.8044.2548.5348.17
70047.3247.9348.2255.3351.20
80053.4353.4355.7062.3462.05
Table 4. The improvement percentage in makespan under different number of tasks.
Table 4. The improvement percentage in makespan under different number of tasks.
Algorithm100200300400500600700800
IMMPA2.60%7.38%4.76%2.14%1.62%0.07%1.28%0%
MRFOSSA16.27%21.06%17.69%15.38%12.59%13.93%14.48%14.30%
HHO11.56%9.99%8.64%4.41%5.96%5.60%1.86%4.08%
WOA31.37%27.05%17.59%12.40%11.37%13.28%7.58%13.89%
Table 5. Comparison based on TEC under different number of tasks.
Table 5. Comparison based on TEC under different number of tasks.
TasksDAMPAIMMPAHHOMRFOSSAWOA
10024002424291531373591
20051085453575463566834
30076258037832192028954
40010,12210,29410,38211,92211,378
50012,48412,74113,23614,43314,280
60014,93414,94615,75316,99316,674
70017,16517,24517,49019,41318,185
80019,28619,71919,98921,85921,316
Table 6. The improvement percentage in TEC under different number of tasks.
Table 6. The improvement percentage in TEC under different number of tasks.
Algorithm100200300400500600700800
IMMPA0.99%6.33%5.12%1.67%2.01%0.07%0.46%0.11%
MRFOSSA23.47%19.63%17.13%15.10%13.50%12.11%11.57%11.76%
HHO17.64%11.22%8.35%2.50%5.68%5.19%1.85%3.51%
WOA33.14%25.25%14.83%11.04%12.57%10.43%5.60%9.52%
Table 7. The improvement percentage in makespan under different numbers of VMs.
Table 7. The improvement percentage in makespan under different numbers of VMs.
Algorithm5060708090100110120130Ave
IMMPA10.48%5.53%7.07%4.00%4.64%6.09%6.33%8.50%8.68%6.81%
MRFOSSA41.76%35.60%36.11%36.33%31.64%33.12%33.00%34.33%27.27%34.35%
HHO30.43%28.57%26.53%24.79%19.27%24.79%28.89%31.38%27.63%26.92%
WOA48.77%50.94%54.63%55.74%43.26%46.33%51.12%48.65%55.94%50.60%
Table 8. The improvement percentage in TEC under different numbers of VMs.
Table 8. The improvement percentage in TEC under different numbers of VMs.
Algorithm5060708090100110120130Ave
IMMPA6.97%2.40%8.39%11.11%3.77%6.10%9.08%4.15%4.69%6.30%
MRFOSSA40.32%36.14%40.76%43.65%36.02%37.61%40.16%37.41%35.37%38.60%
HHO30.82%23.23%31.51%30.60%21.41%27.46%29.33%25.65%23.34%27.04%
WOA48.46%33.39%54.38%49.19%41.11%49.75%53.25%44.35%53.61%47.50%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, D.; Zhang, Y. Diversity-Aware Marine Predators Algorithm for Task Scheduling in Cloud Computing. Entropy 2023, 25, 285. https://doi.org/10.3390/e25020285

AMA Style

Chen D, Zhang Y. Diversity-Aware Marine Predators Algorithm for Task Scheduling in Cloud Computing. Entropy. 2023; 25(2):285. https://doi.org/10.3390/e25020285

Chicago/Turabian Style

Chen, Dujing, and Yanyan Zhang. 2023. "Diversity-Aware Marine Predators Algorithm for Task Scheduling in Cloud Computing" Entropy 25, no. 2: 285. https://doi.org/10.3390/e25020285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop