Next Article in Journal
A Nonlinear Structure of a Chemical Reaction Model and Numerical Modeling with the New Aspect of Existence and Uniqueness
Next Article in Special Issue
A Two-State Dynamic Decomposition-Based Evolutionary Algorithm for Handling Many-Objective Optimization Problems
Previous Article in Journal
Data Analysis and Domain Knowledge for Strategic Competencies Using Business Intelligence and Analytics
Previous Article in Special Issue
Moth Search: Variants, Hybrids, and Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Knowledge-Based Evolutionary Optimizing Makespan and Cost for Cloud Workflows

1
School of Electronic Engineering, Xidian University, Xi’an 710071, China
2
Inner Mongolia Institute of Dynamical Machinery, Hohhot 010010, China
3
School of Management, Hunan Institute of Engineering, Xiangtan 411104, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(1), 38; https://doi.org/10.3390/math11010038
Submission received: 1 November 2022 / Revised: 6 December 2022 / Accepted: 19 December 2022 / Published: 22 December 2022
(This article belongs to the Special Issue Biologically Inspired Computing)

Abstract

:
Workflow scheduling is essential to simultaneously optimize the makespan and economic cost for cloud services and has attracted intensive interest. Most of the existing multi-objective cloud workflow scheduling algorithms regard the focused problems as black-boxes and design evolutionary operators to perform random searches, which are inefficient in dealing with the elasticity and heterogeneity of cloud resources as well as complex workflow structures. This study explores the characteristics of cloud resources and workflow structures to design a knowledge-based evolutionary optimization operator, named KEOO, with two novel features. First, we develop a task consolidation mechanism to reduce the number of cloud resources used, reducing the economic cost of workflow execution without delaying its finish time. Then, we develop a critical task adjustment mechanism to selectively move the critical predecessors of some tasks to the same resources to eliminate the data transmission overhead between them, striving to improve the economic cost and finish time simultaneously. At last, we embed the proposed KEOO into four classical multi-objective algorithms, i.e., NSGA-II, HypE, MOEA/D, and RVEA, forming four variants: KEOO-NSGA-II, KEOO-HypE, KEOO-MOEA/D, and KEOO-RVEA, for comparative experiments. The comparison results demonstrate the effectiveness of the KEOO in improving these four algorithms in solving cloud workflow scheduling problems.

1. Introduction

Big data processing applications from various domains, e.g., the Earthquake and Internet of Things, can be divided into a series of phases. In addition, tasks belonging to different phases have complex data dependencies. These applications are commonly described as workflows [1,2]. Due to the substantial computation and data transmission requirements, executing these workflows often requires massive high-performance infrastructures. With the benefits of pay-per-use, elasticity, scalability, high reliability, and flexibility, cloud computing has become an increasingly attractive choice for enterprises to process their workflow applications by alleviating the burden of building, operating, and maintaining infrastructure [3,4,5].
Scheduling workflows in clouds, which determines the mappings from tasks to resources and the task order on each resource, is of paramount importance to optimize the execution makespan and economic cost, satisfying the quality of service for cloud users and earning more profits for cloud providers [6]. Since the cloud workflow scheduling problem is NP-complete [7,8] and involves multiple conflicting objectives [9], many studies choose evolutionary optimization techniques to obtain satisfactory solutions within an acceptable time [10]. Most existing studies adopt multi-objective optimization for cloud workflow scheduling by designing new selection or reproduction operators.
Some studies have tried to design selection operators for multi-objective cloud workflow scheduling. For instance, Zhou et al. [11] combined a fuzzy-dominance-sort-based selection operator with the earliest-finish-time-based reproduction operator to balance makespan and economic cost for workflow execution in clouds. Kumar et al. [12] integrated the entropy weight approach into a multi-criteria decision-making technique to optimize makespan, economic cost, energy consumption, and reliability. Ye et al. [13] enhanced knee point-driven evolutionary algorithm to balance makespan, the average execution time of all workflow tasks, reliability, and economic cost of workflow execution. Pham et al. [14] considered the volatility of spot cloud instances and employed a multi-objective evolutionary algorithm to balance makespan and economic cost for workflow execution in clouds.
Compared with the selection operators, designing problem-specific reproduction operators attracts more interest. Up to now, considerable efforts have been devoted to developing new reproduction operators for multi-objective workflow scheduling in cloud computing [10,15,16]. At first, some works [17,18,19] replaced the reproduction operator of the multi-objective optimization framework with the list-based heuristic rule for cloud workflow scheduling. For instance, Durillo and Prodan [17] improved a task list-based heuristic rule to obtain a series of intermediate non-dominated solutions for each workflow task, and employed the fast non-dominated sorting-based approach [20] to maintain a predefined number of non-dominated solutions. Wu et al. [21] designed a task list-based optimization rule and a preference weight evolutionary strategy to balance the economic cost and makespan of cloud workflows. Although the heuristics-based multi-objective optimization algorithms pose low time overheads and are effective in specific scenarios, their global search capacity is poor, especially in the face of complex and diverse workflows.
Then, bio-inspired optimization techniques, such as ant colony optimization [22], particle swarm optimization [23,24], artificial neural network [25], genetic algorithm [26,27], and grey wolf optimization [28], were adopted to improve the capacity of reproduction operators in multi-objective evolutionary algorithms. For instance, Zhu et al. [9] improved the multi-objective evolutionary algorithm with problem-specific encoding, population initialization, and reproduction operators to optimize both the makespan and economic cost of cloud workflows. Chen et al. [22] designed a particle swarm optimization algorithm with two colonies to balance the makespan and economic cost of workflow execution in clouds. Ismayilov et al. [25] incorporated an artificial neural network into the NSGA-II to optimize six objectives of workflow execution in cloud computing. Wang et al. [24] embedded idle time gap-based strategies into particle swarm optimization to optimize the economic cost of workflows. In addition, some works integrated heuristic rules and bio-inspired algorithms to reproduce the new population. For instance, Choudhary et al. [29] suggested a hybridization of gravitational search algorithm and heterogeneous-earliest-finish-time for bi-objective scheduling cloud workflows. Mohammadzadeh et al. [30] suggested a hybridization of the antlion and grasshopper optimization algorithm to optimize makespan, economic cost, energy consumption, and throughput of workflow execution in clouds. These existing multi-objective cloud workflow scheduling algorithms often regard the focused problems as black-boxes, and search the solutions in a random way, resulting in low efficiency in dealing with the elasticity and heterogeneity of cloud resources as well as complex workflow structures.
By analyzing these existing works, we can derive that workflow scheduling challenges the randomness of evolutionary search from two aspects. On the one hand, the elasticity, heterogeneity, and on-demand use of cloud computing resources provide a vast number of candidate resources for each workflow task, meaning that the search space for scheduling a cloud workflow is expanding explosively. It is inefficient for the multi-objective evolutionary algorithms to search explosive growth space randomly. On the other hand, due to data dependency among workflow tasks, randomly adjusting a task’s execution scheme often successively affects a series of tasks, including its successor tasks and those tasks being executed after these tasks and their successor tasks.
The current challenges of scheduling cloud workflows motivate us to explore the knowledge of cloud resources and workflow structures to design an effective multi-objective evolutionary optimization algorithm. More specifically, we explore the heterogeneity, pay-as-you-go, and elasticity of cloud resources to design a consolidation mechanism to merge workflow tasks on different cloud resources without delaying any tasks. This way helps reduce the economic cost by reducing the number of cloud resources used and improve search efficiency by shrinking the set of candidate cloud resources. Moreover, the knowledge that data transmission overheads among tasks on the same resource are negligible motivates us to design a critical task adjustment mechanism. It selectively moves the critical predecessors of some tasks to the same resources to eliminate the data transmission overhead between them, striving to improve the finish time and economic cost simultaneously. At last, based on real-world workflows and cloud platforms, we conduct comparative experiments to demonstrate that the proposed approach is capable of improving the performance of multi-objective evolutionary algorithms in solving cloud workflow scheduling problems.
This paper is organized as follows. Section 2 mathematically formulates the multi-objective workflow scheduling problem. Section 3 describes the proposed KEOO, followed by experimental verifications in Section 4. Section 5 concludes this paper.

2. Problem Formulation

This section provides the models for workflow and cloud resource, then formulates the multi-objective cloud workflow scheduling problem.

2.1. Workflow Model

Generally, the workflow structure is modeled as a Directed Acyclic Graph (DAG), in which the nodes and directed edges denote the tasks and the data dependencies among the tasks, respectively. In detail, the DAG model of a workflow is formulated as Ψ = { T , D } , where T = { t 1 , t 2 , , t n } is the set of nodes representing the workflow tasks, and D T × T is the set of edges representing data dependencies among the tasks. The existence of edge d i , j D means that the start of task t j requires the output result of task t i as input. Then, task t i is regarded as an immediate predecessor of task t j , and t j is regarded as an immediate successor of t i . For a task t i , the set of all its immediate predecessors is represented as P ( t i ) , while the set of all its immediate successors is represented as S ( t i ) .
Figure 1 provides an intuitive example of a DAG model for a workflow having seven tasks, i.e., T = { t 1 , t 2 , , t 7 } . The edge d 1 , 2 represents the data dependency between t 1 and t 2 , meaning that the start of task t 2 needs to wait for the output result of task t 1 . As can be seen in Figure 1, for task t 7 , the set of its immediate predecessors is P ( t 7 ) = { t 5 , t 6 } , and the set of t 6 ’s immediate successors is S ( t 6 ) = { t 7 } .

2.2. Cloud Resource Model

Infrastructure as a Service (IaaS) is one popular cloud paradigm, where cloud providers provide unlimited cloud resources with various types [31]. Different resource types differ in price and performance configurations, such as CPU frequency, network bandwidth, memory, and storage size. Given that the cloud platform provides m types of resources, we describe all these resource types as Γ = { 1 , 2 , , m } , where τ Γ corresponds to the τ -th resource type. For a resource type τ , its price and configurations are respectively represented as p r ( τ ) and c o n ( τ ) . Then, a resource instance of type τ in a cloud platform can be modeled as r k τ = { k , p r ( τ ) , c o n ( τ ) } , where k denotes the index of this resource instance.
We refer to well-known cloud providers, e.g., Amazon EC2 and Alibaba Cloud ECS, and follow their resource charging mode of pay-as-you-use. This way, each user can rent cloud instances on-demand and pay for the used instances based on the real usage time. Generally, the cloud providers charge for resource instances according to the number of charging periods and round up the partial time of a period to one more period. If the period length is one hour, the number of charging periods for 60.5 min is two.
The network structure among resource instances in clouds is often heterogeneous and intricate. Since this paper focuses on scheduling workflow tasks, we simplify the underlying network structure and assume that all resource instances are interconnected. The symbol b k , l is employed to denote the communication bandwidth between resource instance r k τ and r l τ . When two data-dependent tasks are executed on the same resource instance, they will use the same storage space, and there is no data transmission through the network. Then, the data transmission overhead between these two tasks can be negligible [9,32].

2.3. Multi-Objective Scheduling Cloud Workflows

Since cloud resource instances are elastic, we construct a resource pool based on the workflow’s most used resource instances. We use p to represent the maximum parallelism of the workflow, and then the resource pool contains p instances of each type. That is to say, the resource pool can be detailed as
R = { r 1 1 , r 2 1 , , r p 1 , r p + 1 2 , r p + 1 2 , , r 2 · p 2 , , r m · p m } .
This paper defines the decision vector x = { x 1 , x 2 , , x n } as the mappings from workflow tasks to resource instances, where the i-th decision variable x i corresponds to the i-th workflow task and its value indicates the index of this task mapping resource instance. Then, the value of each decision variable is one of the elements of the set { 1 , 2 , , m · p } .
For a decision vector, we assume that the workflow task t i is mapped to resource instance r k τ . This task’s start time s t i , k is the maximum time to collect the output results from all its predecessors and the available mapped resource instance.
On resource instance r k τ , the task set ahead of task t i is described as:
B i = { t p | O ( t p ) < O ( t i ) } ,
where O ( t p ) denotes the order number of task t p on the resource instance r k τ .
Then, the start time s t i , k of workflow task t i on the mapped resource instance r k τ is calculated as follows:
s t i , k = max { max t b B i f t b , k , max t p P ( t i ) { f t p , + d t p , i } } ,
where f t b , k denotes the finish time of task t b on resource instance r k τ , f t p , denotes the finish time of task t p on its mapped resource instance, and d t p , i denotes the data transmission time from t p to t i .
Before scheduling, the execution time e t i , k of workflow task t i on resource instance r k τ can be estimated by the computation amount of the task and the performance configuration of the resource instance. Then, the relationships among s t i , k , e t i , k , and f t i , k can be described as follows:
f t i , k = s t i , k + e t i , k .
The data dependencies among workflow tasks mean that a task t i cannot start execution before receiving the output data from all its predecessors, which creates the following constraint:
s t i , k max t p P ( t i ) { f t p , r ( t p ) + I { r ( t i ) r ( t p ) } × w ( e p , i ) b w } , t i T ,
where I { · } is an indicator function. If t p and t i are mapped to different resources, I { · } is 1; otherwise, it is 0. The indicator function is employed to reflect the fact that once two dependent tasks are executed by the same resource, the data transmission overhead between these two tasks is negligible and assumed to be zero. b w denotes the bandwidth.
Given a decision vector, the set of all tasks mapped to resource instance r k τ can be described as:
T k = { t i | x i = k , i { 1 , 2 , , n } } .
With the mapped task set T k , the startup time u t k and shutdown time n t k of resource instance r k τ can be calculated as follows:
u t k = min t i T k { s t i , k max t p P ( t i ) d t p , i } , n t k = max t i T k { f t i , k + max t s S ( t i ) d t i , s } .
With the formulation above, we formulate the first optimization objective, i.e., minimizing the economic cost, as follows:
Min   f 1 ( x ) = r k τ R p r ( τ ) × n t k u t k C ,
where C denotes the length of charging period for resource instances.
The second optimization objective is to minimize the makespan of the workflow, which refers to the maximum finish time of all the workflow tasks. We formulate this optimization objective as follows:
Min f 2 ( x ) = max t i T f t i , .
To summarize, the model for multi-objective scheduling cloud workflows can be formulated as follows:
Min f ( x ) = [ f 1 ( x ) , f 2 ( x ) ] , S.t. x { 1 , 2 , , m · p } n , s t i , k max t p P ( t i ) { f t p , r ( t p ) + I { r ( t i ) r ( t p ) } × w ( e p , i ) b w } , t i T .
To improve readability, we take the workflow in Figure 1 as an example to visually illustrate the decision variable and the calculation of the corresponding objective vector. Assuming that the resource set is R = { r 1 , r 2 , , r 5 } , one decision variable of the workflow with seven tasks in Figure 1 is x = { 2 , 1 , 3 , 2 , 4 , 5 , 2 } . The configurations of the five cloud resources and the execution time from tasks to resources are summarized in Table 1. The data transmission time among workflow tasks is given in Table 2. Based on the above assumptions, Figure 2 illustrates the Gantt chart of the schedule. Then, we can calculate each workflow task’s start and finish time, which is given in Table 3.
According to the charging mode of cloud resources, i.e., the partial time of a charging period is rounded up to one, the charging periods of the five resources are 1, 2, 1, 1, and 1, respectively. Then, the execution cost of the workflow is 0.025 × 1 + 0.032 × 2 + 0.025 × 1 + 0.025 × 1 + 0.032 × 1 = 0.171 dollars, which corresponds to the first optimization objective. Besides, the makespan of a workflow refers to the maximum finish time of all the tasks. We can obtain this optimization objective as max { 5.0 , 13.0 , 18.0 , 20.0 , 38.0 , 45.0 , 66.0 } = 66.0 min.
Pareto dominance is commonly used to compare solutions with multiple conflicting objectives [33,34].
Pareto-Dominance: Suppose x 1 and x 2 are two feasible solutions for the cloud workflow scheduling. x 1 is regarded to Pareto dominate x 2 (expressed as x 1 x 2 ) if and only if the two objectives of x 1 is no larger than that of x 2 (i.e., f j ( x 1 ) f j ( x 2 ) , j { 1 , 2 } ) and x 1 is less than x 2 on at least one objective (i.e., f j ( x 1 ) < f j ( x 2 ) , j { 1 , 2 } ).
Pareto-optimal Solution: A solution x * { 1 , 2 , , m · p } n is generally defined as Pareto-optimal when there exists no feasible solution dominating it.
Pareto-optimal Set/Front: All the Pareto-optimal solutions are defined as Pareto-Set (PS) in the decision space and Pareto-Front (PF) in the objective space.

3. Algorithm Design

The framework of traditional multi-objective evolutionary algorithms includes initialization, reproduction operator, and selection operator [35]. The proposed KEOO contributes to strengthening reproduction operators’ search capability by exploring the knowledge of cloud resources and workflow structures. It incorporates a task consolidation mechanism to efficiently search explosive growth solution space caused by the heterogeneity and elasticity of cloud resources, and a critical task adjustment mechanism to handle the complex workflow structures. Algorithm 1 provides the overall framework of a multi-objective evolutionary algorithm embracing the proposed KEOO.
Algorithm 1: The main framework of KEOO
Mathematics 11 00038 i001
As illustrated in Algorithm 1, the inputs of the proposed KEOO are the multi-objective cloud workflow scheduling problem and the population size. Once the KEOO reaches the termination condition, it outputs an up-to-date population.
In the initialization stage, one population is generated randomly (Line 1). Then, the KEOO iterates the reproduction and environmental selection stages until the termination condition is reached. During the reproduction stage, either the task consolidation mechanism (Line 5) or the critical task adjustment mechanism (Line 7) is triggered to reproduce an offspring population. Because the task consolidation mechanism is a heuristic rule and lacks global search capability, this mechanism is designed to further optimize the offspring solutions generated by traditional evolutionary algorithms (Line 4), rather than being used alone. Because the focused cloud workflow scheduling problem in this paper only optimizes two objectives of makespan and economic cost, many classical environmental selection operators are competent. Thus, this paper does not design a new selection operator but directly employs existing ones, such as NSGA-II [20], HypE [36], MOEA/D [37], and RVEA [38].
Algorithm 2 gives the pseudo-code of the task consolidation mechanism, which explores the heterogeneity, pay-as-you-go, and elasticity of cloud resources to consolidate workflow tasks to as few cloud resources as possible. Then, the number of leased resources can be reduced to achieve the goal of reducing the economic cost.
Algorithm 2: Function TaskConsolidation()
Mathematics 11 00038 i002
As illustrated in Algorithm 2, the function TaskConsolidation() consists of two phases: resource classification and task consolidation. To avoid affecting the task execution, the function only merges tasks on the resources with the same configurations. Then, during the phase of resource classification, all the resources are classified into a series of groups (Lines 1–11). At first, this function sorts all the resources according to their CPU configurations (Line 1). The symbol k is used to record the group number, initialized as 1 (Line 2). The set G k is used to record the resources belonging to the k-th group (Line 3). The operator C O N F ( · ) means to obtain the configuration information of a cloud resource. Next, each resource is checked (Line 5). If its configuration is consistent with the recorded one, it will be added to the current group (Line 7). Otherwise, a new group is created (Lines 9–10), and the recorded configuration is updated (Line 11).
After resource classification, the function TaskConsolidation() enters the task consolidation phase. It sorts a group of cloud resources in ascending order according to the maximum finish time of the mapped tasks (Line 13). Then, the first resource in this group is selected (Line 14), and the completion time of this resource is recorded (Line 15). If the earliest start time of all tasks mapped to a cloud resource is greater than the record time (Line 17), all tasks on this resource are consolidated to the selected resource (Lines 18–20). Otherwise, a cloud resource is reselected and the record time is updated (Lines 22–23).
To illustrate the function TaskConsolidation() clearly, Figure 3 gives a visual example based on the schedule in Figure 2. According to the configurations of the five cloud resources in Table 1, they can be divided into two groups, i.e., G 1 = { r 1 , r 3 , r 4 } and G 2 = { r 2 , r 5 } . The three cloud resources in the first group will be sorted as r 1 , r 3 , and r 4 . The earliest start time of tasks on resource r 3 is not greater than the maximum finish time of tasks on resource r 1 , so the task consolidation requirements are not met. On the contrary, the earliest start time of tasks on resource r 4 is greater than the maximum finish time of task on resource r 1 , then the task on r 4 is consolidated to r 1 , as illustrated in Figure 3. By comparing Figure 2 and Figure 3, we can observe that resource r 4 can be released, reducing the economic cost of this resource. After task consolidation, according to the charging mode of cloud resources, the charging period of resource r 3 is still 1, and the economic cost remains unchanged. This means that the proposed task consolidation mechanism helps reduce the economic cost without affecting the task execution.
The start time of a workflow task is constrained by the completion time of collecting the output results from all its predecessors. Besides, if two data-dependent tasks run on the same cloud resource, the data transmission overhead between them is negligible. The above facts motivate us to adjust some tasks to the same resources to simultaneously improve the finish time and economic cost of workflow execution, as illustrated in Algorithm 3.
Algorithm 3: Function CriticalTaskReschedule()
Mathematics 11 00038 i003
The Function (CriticalTaskReschedule) consists of two parts: identification and adjustment of critical predecessor tasks. For a workflow task, its start time is limited by all its predecessors, and its critical predecessor is defined as the one with the latest result arrival. This function first identifies the critical predecessor for each task (Lines 1–7). The value of the i-th element in C I denotes the index of the critical predecessor for the i-th task.
Due to the complex workflow structure, a task may be the critical predecessor for multiple tasks, and there may be a critical predecessor. To alleviate the offset of critical task adjustments, adjusting the critical predecessor of a task to its mapped resource should satisfy the following three conditions: (1) this task has a critical predecessor, i.e., p 0 ; (2) the critical predecessor has not been adjusted, i.e., T a g s ( p ) = = 0 ; (3) this task has not been adjusted as a critical predecessor, i.e., T a g s ( i ) = = 0 .
Take the scheduling scheme in Figure 3 as an example. According to the definition of a critical predecessor in Function (CriticalTaskReschedule), t 1 is the critical predecessor of t 2 , and t 6 is the critical predecessor of t 7 . Adjusting task t 1 from resource r 2 to r 1 , as shown in Figure 4a, the data transmission overhead between task t 1 and t 2 can be eliminated to advance the start time of those tasks including t 2 and its successor tasks. Added to that, adjusting task t 6 from cloud resource r 5 to r 2 , as shown in Figure 4b, advances the start/finish time of task t 7 . This helps to simultaneously reduce the makespan and economic cost of executing the workflow.

4. Experimental Studies

The main work of this paper is to propose a task consolidation and a critical task adjustment mechanism, thereby improving the quality of reproducing new populations. The proposal does not involve environmental selection. We employ the proposed two mechanisms to replace the reproduction operators of four classical multi-objective evolutionary algorithms, i.e., NSGA-II [20], HypE [36], MOEA/D [37], and RVEA [38], forming four variants, namely, KEOO-NSGA-II, KEOO-HypE, KEOO-MOEA/D, and KEOO-RVEA. Then, the four variants are compared with their original versions to verify the effectiveness of the two proposed mechanisms.

4.1. Experimental Setting

We perform the comparison experiments based on five different types of resources provided by the Amazon EC2 cloud platform. These five types of resource instances are t3.nano, t3.micro, t3.small, t3.medium, t3.medium, and t3.large. The main parameters of the five resource types are summarized in Table 4. The length of a charging period is set to 60 seconds, and the bandwidth among resource instances is set to 5.0 Gbps.
The five types of workflows from different application domains, i.e., Montage (Astronomy), Epigenomics (Biology), Inspiral (Gravitational physics), CyberShake (Earthquake), and Sipht (Bioinformatics), have been widely used in evaluating cloud workflow scheduling algorithms. For each type of workflow, we select three workflow instances with about 30, 50, and 100 tasks in the experiments. Besides, the DAG diagrams of the workflow instances with around 30 tasks are illustrated in Figure 5. It is clear that these workflow instances cover various complicated structures, including in-tree, out-tree, fork-join, pipeline, and mixture. For more details on these workflows, please refer to Pegasus repository.
Hypervolume [39] metric is designed to measure the quality of a population concerning both convergence and diversity, and has been frequently used to evaluate the performance of multi-objective evolutionary algorithms. Assume that r = { r 1 , r 2 } is a reference point. The hypervolume value of a population P, corresponding to the volume between the reference point and the objective vectors of the solutions in P, can be calculated as follows.
H V ( P ) = L ( p P [ f 1 ( p ) , r 1 ] × [ f 2 ( p ) , r 2 ] ) ,
where L ( ) represents the Lebesgue measure.
The population size of the eight algorithms is set as 120. The maximum number of fitness evaluations is set as the termination condition for all the five algorithms and is set as n × 5 × 10 3 , where n is the number of decision variables.
The eight algorithms are independently repeated 31 times on each workflow instance to mitigate the impact of random factors. We run all the experiments on a PC with two Cores i7-6500U CPU @ 2.50 GHz 2.59 GHz, 8.00 GB RAM, Windows 10.

4.2. Comparison Results

Table 5 summarizes the average and standard deviation (in brackets) of the hypervolume values for the eight algorithms, i.e., NSGA-II, KEOO-NSGA-II, HypE, KEOO-HypE, MOEA/D, KEOO-MOEA/D, RVEA, and KEOO-RVEA, in scheduling the 15 workflow instances to cloud resources. For each workflow instance, the largest hypervolume value among the eight algorithms is highlighted using a gray background.
Besides, we resort to the Wilcoxon rank-sum test with a significance of 0.1 to differentiate the significant difference between each variant and its source baseline. The signs −, +, and ≈ indicate that the variant is significantly inferior to, superior to, and similar to the corresponding baseline, respectively.
From Table 5, we can observe that except for Inspiral 50 and Cybershake 100, the results marked with a gray background on the other 13 workflow instances are either KEOO-NSGA-II or KEOO-HypE. Specifically, KEOO-NSGA-II significantly improves the hypervolume values of NSGA-II on 10 out of 15 workflow instances, and KEOO-HypE improves its original version on 10 out of 15 workflow instances. The KEOO-NSGA-II and KEOO-HypE are variants of NSGA-II or HypEA by integrating the proposed task consolidation and critical task adjustment mechanisms. The comparison results demonstrate that the proposed mechanisms can effectively improve the performance of the NSGA-II or HypEA.
MOEA/D and RVEA are representative algorithms of two branches of multi-objective evolutionary algorithms based on decomposition. One branch transforms the multi-objective optimization problem into a set of single-objective subproblems using a set of weight vectors, while the other branch divides the multidimensional objective space into a series of multi-objective subspaces using a set of weight vectors. Compared with NSGA-II or HypEA, these two algorithms and their variants pose poor performance in solving multi-objective optimization problems. The key reason is that the value ranges of the two objectives of the focused problems are far from each other, typically badly-scaled problems. Then, the intersection points between the Pareto-optimal fronts of the focused problems and the weight vectors are unevenly distributed, which weakens the performance of these multi-objective evolutionary algorithms based on decomposition. It can be seen from Table 5 that for more than half of the workflow instances, variant KEOO-RVEA has higher hypervolume values than algorithm RVEA. KEOO-MOEA/D and MOEA/D have similar comparison results. These comparison results demonstrate that the proposed task consolidation and critical task adjustment mechanisms can effectively improve the performance of multi-objective evolutionary algorithms based on decomposition in solving multi-objective cloud workflow scheduling problems.
From Table 5, the difference in hypervolume values obtained by the eight algorithms is not apparent on the same workflow instances. This is because the reference point for calculating the hypervolume value is set based on the nadir point of the initial population, which is often far from the output populations of these algorithms. Thus, each algorithm can obtain a very large hypervolume value, resulting in slight differences.
It can be seen from Figure 5 that among the five types of workflows, the Montage workflow has the most complex structure. From Table 5, we also observe that on the four workflow instances derived from this application, the proposed task consolidation and critical task adjustment mechanisms can greatly improve the hypervolume values of the four classical multi-objective evolutionary algorithms. These improvements demonstrate the effectiveness of the proposed mechanisms in dealing with complex workflow structures.
To intuitively compare the convergence and diversity of the eight multi-objective workflow scheduling algorithms, Figure 6 illustrates the distribution of their output populations on workflow instances Inspiral, Epigenomics, Montage, CyberShake, and Sipht with around 30 tasks as well as Sipht with around 100 tasks.
In Figure 6, the populations obtained by the algorithms using the proposed KEOO are shown in red, and different algorithms are distinguished by different icons. The first impression of the six sub-figures is that the red icons are distributed at the forefront. This indicates that the proposed KEOO improves the existing multi-objective evolutionary algorithms on both makespan and economic cost. These comparison results are consistent with those in Table 5. The population distribution of the variant with the proposed KEOO is basically consistent with that of the original algorithm. This is because they all use the same environmental selection operator. The advantage of the four variants is that they employ the task consolidation mechanism to reduce economic cost, and the critical task adjustment mechanism to reduce makespan and economic cost simultaneously.
On workflow instances Inspiral 30 and Epigenomics 24, although the solutions of the algorithms embedded with the proposed mechanisms distribute at the forefront, they are similar to the solutions of their corresponding original algorithms. The reason can be attributed to the fact that these two workflow instances can be divided into multiple independent simple pipelines, and the corresponding optimization problems are relatively easy. The traditional multi-objective evolutionary algorithms can solve these simple problems well, leaving limited room for improvement.
On workflow instances Montage 25 and Cybershake 30, solutions obtained by KEOO-NSGA-II and KEOO-HypE are obviously separated from the solutions of their original algorithms. The reason is that these two workflow instances are relatively complex. When scheduling such workflow instances, the start and completion time of different resources vary greatly, leaving room for the task consolidation mechanism to play the advantage of reducing economic cost. A distinctive feature of workflow instance Cybershake 30 is that one of its tasks has a large number of predecessors, and these predecessors no longer have predecessors. Besides, the tasks in Cybershake 30 need to transfer a large amount of data, resulting in many large idle time gaps between tasks. The proposed critical task adjustment mechanism can adjust some critical tasks to idle time gaps to simultaneously reduce the makespan and economic cost.
By comparing Figure 6e,f, we can see that the advantage of algorithms embedded with the proposed mechanisms on Sipht 100 is much more obvious than on Sipht 30. This comparison result illustrates that the proposed mechanisms help accelerate population convergence and effectively deal with the explosively growing search space.
To intuitively demonstrate the advantages of the proposed mechanisms in accelerating the convergence of multi-objective evolutionary algorithms, Figure 7 illustrates the changes in the hypervolume values of eight algorithms with the evolution process. We can observe that the hypervolume values of all eight algorithms rise rapidly in the early stage of the evolutionary search. The reason is that the cloud resource pool is large enough, meaning a wide range of decision variable values. The quality of random initial solutions is poor, and it is relatively easy for evolutionary operators to push them toward the Pareto-optimal fronts. Concerning the hypervolume metric, the growth rate of the algorithms embedded with the proposed mechanisms is faster than that of their original algorithms throughout the search process. These results demonstrate that the proposed mechanisms are capable of facilitating the convergence rate.
For the execution time, the comparison results of the eight algorithms are shown in Table 6. Obviously, the execution time of the variants embedded in the proposed mechanism in this paper is greater than that of the original algorithms. The increased execution cost mainly comes from the proposed mechanism. Fortunately, the execution time of the four variants is about the same order of magnitude as that of the original algorithms, and these variants can obtain populations with better convergence and diversity.

5. Conclusions

This paper mathematically formulates the workflow scheduling problem in cloud computing as a bi-objective optimization problem. Then, it explores the knowledge of cloud resources and workflow structures to design a task consolidation mechanism to reduce economic cost by reducing the number of cloud resources; and a critical task-based search operator to selectively move the critical precursors of some tasks to the same resources to eliminate the data transmission overhead between them, striving to improve the economic cost and completion time simultaneously. Based on real-world workflows and cloud platforms, the proposed mechanisms are embedded into four classical multi-objective evolutionary algorithms to verify their effectiveness in improving population convergence and diversity. One disadvantage of the proposed mechanisms is that they bring extra time overhead to the classical algorithms.
Cloud workflow scheduling is a representative grey-box problem, and it is interesting to further mine the knowledge of the workflows and cloud resources to derive efficient scheduling algorithms. Through experimental analysis, we can see that the time overhead of evolutionary optimization can not be ignored. Therefore, another potential direction is to design an effective parallel evolutionary framework to shorten the time overhead to support cloud workflow scheduling in real-time and uncertain situations [40].

Author Contributions

Conceptualization, L.X. and J.L.; methodology, J.L.; software, L.X. and R.W.; validation, L.X., R.W., J.C. and J.L.; formal analysis, R.W.; investigation, R.W.; resources, J.C.; data curation, J.C.; writing—original draft preparation, L.X.; writing—review and editing, R.W., J.C. and J.L.; visualization, J.C.; supervision, J.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research work is supported by the National Natural Science Foundation of China (61773120), the Special Projects in Key Fields of Universities in Guangdong (2021ZDZX1019).

Data Availability Statement

All data used during the study appear in the submitted article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, H.; Wen, J.; Pedrycz, W.; Wu, G. Big data processing workflows oriented real-time scheduling algorithm using task-duplication in geo-distributed clouds. IEEE Trans. Big Data 2018, 6, 131–144. [Google Scholar] [CrossRef]
  2. Bugingo, E.; Zhang, D.; Chen, Z.; Zheng, W. Towards decomposition based multi-objective workflow scheduling for big data processing in clouds. Clust. Comput. 2021, 24, 115–139. [Google Scholar] [CrossRef]
  3. Molnár, B.; Benczúr, A. The application of directed hyper-graphs for analysis of models of information systems. Mathematics 2022, 10, 759. [Google Scholar] [CrossRef]
  4. Cong, P.; Li, L.; Zhou, J.; Cao, K.; Wei, T.; Chen, M.; Hu, S. Developing user perceived value based pricing models for cloud markets. IEEE Trans. Parallel Distrib. Syst. 2018, 29, 2742–2756. [Google Scholar] [CrossRef]
  5. Jung, A.; Gsell, M.A.; Augustin, C.M.; Plank, G. An integrated workflow for building digital twins of cardiac electromechanics—A multi-fidelity approach for personalising active mechanics. Mathematics 2022, 10, 823. [Google Scholar] [CrossRef]
  6. Farid, M.; Latip, R.; Hussin, M.; Hamid, N.A.W.A. Scheduling scientific workflow using multi-objective algorithm with fuzzy resource utilization in multi-cloud environment. IEEE Access 2020, 8, 24309–24322. [Google Scholar] [CrossRef]
  7. Masdari, M.; ValiKardan, S.; Shahi, Z.; Azar, S.I. Towards workflow scheduling in cloud computing: A comprehensive analysis. J. Netw. Comput. Appl. 2016, 66, 64–82. [Google Scholar] [CrossRef]
  8. Zhang, M.; Li, H.; Liu, L.; Buyya, R. An adaptive multi-objective evolutionary algorithm for constrained workflow scheduling in Clouds. Distrib. Parallel Databases 2018, 36, 339–368. [Google Scholar] [CrossRef]
  9. Zhu, Z.; Zhang, G.; Li, M.; Liu, X. Evolutionary multi-objective workflow scheduling in cloud. IEEE Trans. Parallel Distrib. Syst. 2016, 27, 1344–1357. [Google Scholar] [CrossRef] [Green Version]
  10. Hosseinzadeh, M.; Ghafour, M.Y.; Hama, H.K.; Vo, B.; Khoshnevis, A. Multi-objective task and workflow scheduling approaches in cloud computing: A comprehensive review. J. Grid Comput. 2020, 18, 327–356. [Google Scholar] [CrossRef]
  11. Zhou, X.; Zhang, G.; Sun, J.; Zhou, J.; Wei, T.; Hu, S. Minimizing cost and makespan for workflow scheduling in cloud using fuzzy dominance sort based HEFT. Future Gener. Comput. Syst. 2019, 93, 278–289. [Google Scholar] [CrossRef]
  12. Kumar, M.S.; Tomar, A.; Jana, P.K. Multi-objective workflow scheduling scheme: A multi-criteria decision making approach. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 10789–10808. [Google Scholar] [CrossRef]
  13. Ye, X.; Liu, S.; Yin, Y.; Jin, Y. User-oriented many-objective cloud workflow scheduling based on an improved knee point driven evolutionary algorithm. Knowl. Based Syst. 2017, 135, 113–124. [Google Scholar] [CrossRef]
  14. Pham, T.P.; Fahringer, T. Evolutionary multi-objective workflow scheduling for volatile resources in the cloud. IEEE Trans. Cloud Comput. 2022, 10, 1780–1791. [Google Scholar] [CrossRef]
  15. Rodriguez, M.A.; Buyya, R. A taxonomy and survey on scheduling algorithms for scientific workflows in IaaS cloud computing environments. Concurr. Comput. Pract. Exp. 2017, 29, e4041. [Google Scholar] [CrossRef]
  16. Zhan, Z.H.; Liu, X.F.; Gong, Y.J.; Zhang, J.; Chung, H.S.H.; Li, Y. Cloud computing resource scheduling and a survey of its evolutionary approaches. ACM Comput. Surv. (CSUR) 2015, 47, 1–33. [Google Scholar] [CrossRef] [Green Version]
  17. Durillo, J.J.; Nae, V.; Prodan, R. Multi-objective energy-efficient workflow scheduling using list-based heuristics. Future Gener. Comput. Syst. 2014, 36, 221–236. [Google Scholar] [CrossRef]
  18. Fard, H.M.; Prodan, R.; Fahringer, T. Multi-objective list scheduling of workflow applications in distributed computing infrastructures. J. Parallel Distrib. Comput. 2014, 74, 2152–2165. [Google Scholar] [CrossRef]
  19. Han, P.; Du, C.; Chen, J.; Ling, F.; Du, X. Cost and makespan scheduling of workflows in clouds using list multiobjective optimization technique. J. Syst. Archit. 2021, 112, 101837. [Google Scholar] [CrossRef]
  20. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  21. Wu, Q.; Zhou, M.; Zhu, Q.; Xia, Y.; Wen, J. MOELS: Multiobjective evolutionary list scheduling for cloud workflows. IEEE Trans. Autom. Sci. Eng. 2020, 17, 166–176. [Google Scholar] [CrossRef]
  22. Chen, Z.G.; Zhan, Z.H.; Lin, Y.; Gong, Y.J.; Gu, T.L.; Zhao, F.; Yuan, H.Q.; Chen, X.; Li, Q.; Zhang, J. Multiobjective cloud workflow scheduling: A multiple populations ant colony system approach. IEEE Trans. Cybern. 2019, 49, 2912–2926. [Google Scholar] [CrossRef] [PubMed]
  23. Gupta, R.; Gajera, V.; Jana, P.K. An effective multi-objective workflow scheduling in cloud computing: A PSO based approach. In Proceedings of the 2016 Ninth International Conference on Contemporary Computing, Noida, India, 11–13 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–6. [Google Scholar]
  24. Wang, Y.; Zuo, X. An Effective Cloud Workflow Scheduling Approach Combining PSO and Idle Time Slot-Aware Rules. IEEE/CAA J. Autom. Sin. 2021, 8, 1079–1094. [Google Scholar] [CrossRef]
  25. Ismayilov, G.; Topcuoglu, H.R. Neural network based multi-objective evolutionary algorithm for dynamic workflow scheduling in cloud computing. Future Gener. Comput. Syst. 2020, 102, 307–322. [Google Scholar] [CrossRef]
  26. Hussain, M.; Wei, L.F.; Abbas, F.; Rehman, A.; Ali, M.; Lakhan, A. A multi-objective quantum-inspired genetic algorithm for workflow healthcare application scheduling with hard and soft deadline constraints in hybrid clouds. Appl. Soft Comput. 2022, 128, 109440. [Google Scholar] [CrossRef]
  27. Abbasian-Naghneh, S.; Kalbasi, R. Implementation of ANN and GA on building with PCM at various setpoints, PCM types, and installation locations to boost energy saving and CO2 saving. Eng. Anal. Bound. Elem. 2022, 144, 110–126. [Google Scholar] [CrossRef]
  28. Abed-Alguni, B.H.; Alawad, N.A. Distributed Grey Wolf Optimizer for scheduling of workflow applications in cloud environments. Appl. Soft Comput. 2021, 102, 107113. [Google Scholar] [CrossRef]
  29. Choudhary, A.; Gupta, I.; Singh, V.; Jana, P.K. A GSA based hybrid algorithm for bi-objective workflow scheduling in cloud computing. Future Gener. Comput. Syst. 2018, 83, 14–26. [Google Scholar] [CrossRef]
  30. Mohammadzadeh, A.; Masdari, M.; Gharehchopogh, F.S. Energy and cost-aware workflow scheduling in cloud computing data centers using a multi-objective optimization algorithm. J. Netw. Syst. Manag. 2021, 29, 1–34. [Google Scholar] [CrossRef]
  31. De Maio, V.; Kimovski, D. Multi-objective scheduling of extreme data scientific workflows in Fog. Future Gener. Comput. Syst. 2020, 106, 171–184. [Google Scholar] [CrossRef]
  32. Calheiros, R.N.; Buyya, R. Meeting deadlines of scientific workflows in public clouds with tasks replication. IEEE Trans. Parallel Distrib. Syst. 2014, 25, 1787–1796. [Google Scholar] [CrossRef] [Green Version]
  33. Coello, C.A.C.; Brambila, S.G.; Gamboa, J.F.; Tapia, M.G.C.; Gómez, R.H. Evolutionary multiobjective optimization: Open research areas and some challenges lying ahead. Complex Intell. Syst. 2020, 6, 221–236. [Google Scholar] [CrossRef] [Green Version]
  34. Chen, H.; Cheng, R.; Wen, J.; Li, H.; Weng, J. Solving large-scale many-objective optimization problems by covariance matrix adaptation evolution strategy with scalable small subpopulations. Inf. Sci. 2020, 509, 457–469. [Google Scholar] [CrossRef]
  35. Li, B.; Li, J.; Tang, K.; Yao, X. Many-objective evolutionary algorithms: A survey. ACM Comput. Surv. (CSUR) 2015, 48, 1–35. [Google Scholar] [CrossRef] [Green Version]
  36. Bader, J.; Zitzler, E. HypE: An algorithm for fast hypervolume-based many-objective optimization. Evol. Comput. 2011, 19, 45–76. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Li, H. MOEA/D: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  38. Cheng, R.; Jin, Y.; Olhofer, M.; Sendhoff, B. A reference vector guided evolutionary algorithm for many-objective optimization. IEEE Trans. Evol. Comput. 2016, 20, 773–791. [Google Scholar] [CrossRef] [Green Version]
  39. Zitzler, E.; Thiele, L. Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach. IEEE Trans. Evol. Comput. 1999, 3, 257–271. [Google Scholar] [CrossRef] [Green Version]
  40. Chen, H.; Zhu, X.; Liu, G.; Pedrycz, W. Uncertainty-aware online scheduling for real-time workflows in cloud service environment. IEEE Trans. Serv. Comput. 2021, 14, 1167–1178. [Google Scholar] [CrossRef]
Figure 1. DAG model of a workflow with seven tasks.
Figure 1. DAG model of a workflow with seven tasks.
Mathematics 11 00038 g001
Figure 2. Example of a schedule.
Figure 2. Example of a schedule.
Mathematics 11 00038 g002
Figure 3. Example of consolidating tasks in different resources.
Figure 3. Example of consolidating tasks in different resources.
Mathematics 11 00038 g003
Figure 4. Example of adjusting critical tasks. (a) schedule result of adjusting task t 1 from resource r 2 to r 1 ; (b) schedule result of adjusting task t 6 from resource r 5 to r 2 .
Figure 4. Example of adjusting critical tasks. (a) schedule result of adjusting task t 1 from resource r 2 to r 1 ; (b) schedule result of adjusting task t 6 from resource r 5 to r 2 .
Mathematics 11 00038 g004
Figure 5. DAG diagrams of workflows with about 30 tasks. (a) Montage. (b) Epigenomics. (c) Inspiral. (d) CyberShake. (e) Sipht.
Figure 5. DAG diagrams of workflows with about 30 tasks. (a) Montage. (b) Epigenomics. (c) Inspiral. (d) CyberShake. (e) Sipht.
Mathematics 11 00038 g005
Figure 6. Distributions of populations obtained by the 8 algorithms on Inspiral, Epigenomics, Montage, CyberShake, and Sipht with around 30 tasks as well as Sipht with around 100 tasks. (a) on Inspiral with 30 tasks; (b) on Epigenomics with 24 tasks; (c) on Montage with 25 tasks; (d) on CyberShake with 30 tasks; (e) on Sipht with 30 tasks; (f) on Sipht with 100 tasks.
Figure 6. Distributions of populations obtained by the 8 algorithms on Inspiral, Epigenomics, Montage, CyberShake, and Sipht with around 30 tasks as well as Sipht with around 100 tasks. (a) on Inspiral with 30 tasks; (b) on Epigenomics with 24 tasks; (c) on Montage with 25 tasks; (d) on CyberShake with 30 tasks; (e) on Sipht with 30 tasks; (f) on Sipht with 100 tasks.
Mathematics 11 00038 g006
Figure 7. Change of hypervolume values with the advance of evolution. (a) on Inspiral with 30 tasks; (b) on Epigenomics with 24 tasks; (c) on Montage with 25 tasks; (d) on CyberShake with 30 tasks; (e) on Sipht with 30 tasks; (f) on Sipht with 100 tasks.
Figure 7. Change of hypervolume values with the advance of evolution. (a) on Inspiral with 30 tasks; (b) on Epigenomics with 24 tasks; (c) on Montage with 25 tasks; (d) on CyberShake with 30 tasks; (e) on Sipht with 30 tasks; (f) on Sipht with 100 tasks.
Mathematics 11 00038 g007
Table 1. Examples of task execution time (in minute).
Table 1. Examples of task execution time (in minute).
r 1 r 2 r 3 r 4 r 5
CPU (GHz)2.93.32.92.93.3
Memory Size (GB)4.06.04.04.06.0
Bandwidth (MB/s)20.030.020.020.030.0
Storage (GB)500.01000.0500.0500.01000.0
Price ($/h)0.0250.0320.0250.0250.032
t 1 6.255.06.256.255.0
t 2 5.04.05.05.04.0
t 3 10.08.010.010.08.0
t 4 6.255.06.256.255.0
t 5 15.012.015.015.012.0
t 6 25.020.025.025.020.0
t 7 22.518.022.522.518.0
Table 2. Examples of data transmission time (in minutes) among tasks.
Table 2. Examples of data transmission time (in minutes) among tasks.
t 1 t 2 t 3 t 4 t 5 t 6 t 7
t 1 33
t 2 2
t 3 5
t 4 35
t 5 10
t 6 3
t 7
Table 3. Start time s t ( · ) and finish time f t ( · ) (in minute) of each workflow task.
Table 3. Start time s t ( · ) and finish time f t ( · ) (in minute) of each workflow task.
t 1 t 2 t 3 t 4 t 5 t 6 t 7
s t ( · ) 0.08.08.015.023.025.048.0
f t ( · ) 5.013.018.020.038.045.066.0
Table 4. Parameters for the five types of cloud resources.
Table 4. Parameters for the five types of cloud resources.
TypePrice ($/h)vCPUMemory (GB)
t3.nano0.006220.5
t3.micro0.012521.0
t3.small0.02522.0
t3.medium0.049924.0
t3.large0.099828.0
Table 5. Comparison results for the 8 algorithms on 15 workflows in terms of the hypervolume metric.
Table 5. Comparison results for the 8 algorithms on 15 workflows in terms of the hypervolume metric.
WorkflowsnNSGA-IIKEOO-NSGA-IIHypEKEOO-HypEMOEA/DKEOO-MOEA/DRVEAKEOO-RVEA
Montage257.013 × 10 2 7.113 × 10 2 +6.924 × 10 2 7.142 × 10 2 +6.078 × 10 2 6.253 × 10 2 +6.741 × 10 2 6.777 × 10 2
(2.23 × 10 1 )(1.55 × 10 1 )(1.38 × 10 1 )(8.95 × 10)(4.61 × 10 1 )(2.54 × 10 1 )(2.63 × 10 1 )(2.56 × 10 1 )
501.361 × 10 3 1.376 × 10 3 +1.370 × 10 3 1.392 × 10 3 +1.243 × 10 3 1.025 × 10 3 1.265 × 10 3 1.327 × 10 3 +
(3.16 × 10 1 )(2.99 × 10 1 )(2.92 × 10 1 )(2.38 × 10 1 )(3.10 × 10 1 )(4.35 × 10 2 )(3.68 × 10 1 )(4.59 × 10 1 )
1001.353 × 10 3 1.476 × 10 3 +2.440 × 10 3 2.449 × 10 3 2.304 × 10 3 1.992 × 10 3 2.211 × 10 3 2.449 × 10 3 +
(3.88 × 10 1 )(5.42 × 10 1 )(5.142 × 10 1 )(7.37 × 10 1 )(5.92 × 10 1 )(9.18 × 10 2 )(9.12 × 10 1 )(3.69 × 10 1 )
Epigenomics245.656 × 10 5 5.678 × 10 5 +5.627 × 10 5 5.690 × 10 5 +5.201 × 10 5 5.173 × 10 5 5.586 × 10 5 5.648 × 10 5 +
(2.50 × 10 3 )(2.64 × 10 3 )(2.84 × 10 3 )(2.17 × 10 3 )(1.13 × 10 3 )(1.48 × 10 3 )(3.68 × 10 3 )(2.24 × 10 3 )
461.899 × 10 6 1.915 × 10 6 +1.893 × 10 6 1.909 × 10 6 +1.815 × 10 6 1.784 × 10 6 1.882 × 10 6 1.894 × 10 6 +
(5.96 × 10 3 )(2.04 × 10 3 )(7.10 × 10 3 )(1.03 × 10 4 )(2.68 × 10 4 )(3.91 × 10 4 )(9.60 × 10 3 )(1.21 × 10 4 )
1005.056 × 10 7 5.066 × 10 7 5.006 × 10 7 5.014 × 10 7 4.843 × 10 7 4.674 × 10 7 5.022 × 10 7 4.962 × 10 7
(3.56 × 10 5 )(2.13 × 10 5 )(3.23 × 10 5 )(3.32 × 10 5 )(5.65 × 10 5 )(1.06 × 10 6 )(2.78 × 10 5 )(4.26 × 10 5 )
Inspiral304.078 × 10 4 4.093 × 10 4 4.066 × 10 4 4.080 × 10 4 3.748 × 10 4 3.577 × 10 4 4.016 × 10 4 4.025 × 10 4
(6.49 × 10 2 )(2.57 × 10 2 )(4.93 × 10 2 )(1.64 × 10 2 )(5.04 × 10 2 )(1.70 × 10 3 )(4.21 × 10 2 )(3.40 × 10 2 )
509.466 × 10 4 9.329 × 10 4 9.366 × 10 4 9.354 × 10 4 8.945 × 10 4 8.360 × 10 4 9.338 × 10 4 9.256 × 10 4
(4.98 × 10 2 )(8.17 × 10 2 )(1.10 × 10 3 )(6.47 × 10 2 )(1.33 × 10 3 )(1.88 × 10 3 )(9.01 × 10 2 )(1.02 × 10 3 )
1001.453 × 10 5 1.444 × 10 5 1.426 × 10 5 1.457 × 10 5 +1.375 × 10 5 1.325 × 10 5 1.436 × 10 5 1.420 × 10 5
(1.27 × 10 3 )(2.01 × 10 3 )(2.17 × 10 3 )(1.45 × 10 3 )(4.34 × 10 3 )(3.39 × 10 3 )(1.80 × 10 3 )(2.35 × 10 3 )
CyberShake302.240 × 10 5 2.301 × 10 5 +2.235 × 10 5 2.485 × 10 5 +2.161 × 10 5 2.4110 × 10 5 +2.1770 × 10 5 2.486 × 10 5 +
(8.96 × 10 3 )(4.07 × 10 3 )(8.31 × 10 3 )(4.84 × 10 3 )(1.14 × 10 3 )(5.98 × 10 3 )(1.97 × 10 3 )(4.66 × 10 3 )
503.011 × 10 5 3.081 × 10 5 +3.038 × 10 5 3.065 × 10 5 +2.966 × 10 5 3.051 × 10 5 +3.024 × 10 5 3.014 × 10 5
(2.43 × 10 3 )(5.98 × 10 3 )(1.77 × 10 3 )(5.87 × 10 3 )(1.52 × 10 3 )(4.78 × 10 3 )(2.34 × 10 3 )(4.23 × 10 3 )
1001.971 × 10 6 1.930 × 10 6 1.985 × 10 6 1.938 × 10 6 1.961 × 10 6 1.951 × 10 6 1.968 × 10 6 1.944 × 10 6
(7.59 × 10 4 )(3.65 × 10 4 )(1.12 × 10 4 )(2.12 × 10 4 )(7.81 × 10 4 )(1.45 × 10 4 )(1.26 × 10 4 )(2.33 × 10 4 )
Sipht308.278 × 10 4 8.344 × 10 4 +8.302 × 10 4 8.342 × 10 4 +7.944 × 10 4 8.031 × 10 4 +8.149 × 10 4 8.260 × 10 4 +
(3.79 × 10 2 )(7.04 × 10 1 )(2.74 × 10 2 )(1.20 × 10 2 )(9.73 × 10 2 )(2.65 × 10 2 )(4.93 × 10 2 )(1.15 × 10 2 )
601.991 × 10 5 2.032 × 10 5 +1.992 × 10 5 2.028 × 10 5 +1.860 × 10 5 1.963 × 10 5 +1.938 × 10 5 2.009 × 10 5 +
(1.17 × 10 3 )(2.63 × 10 2 )(1.30 × 10 3 )(4.14 × 10 3 )(4.64 × 10 3 )(1.12 × 10 3 )(1.64 × 10 3 )(4.24 × 10 2 )
1003.427 × 10 5 3.615 × 10 5 +3.461 × 10 5 3.598 × 10 5 +2.996 × 10 5 3.502 × 10 5 +3.292 × 10 5 3.536 × 10 5 +
(6.29 × 10 3 )(8.31 × 10 2 )(4.31 × 10 3 )(1.98 × 10 3 )(1.19 × 10 4 )(3.718 × 10 3 )(5.914 × 10 3 )(1.430 × 10 3 )
Table 6. Execution time (second) of the eight algorithms.
Table 6. Execution time (second) of the eight algorithms.
WorkflowsnNSGA-IIKEOO-NSGA-IIHypEKEOO-HypEMOEA/DKEOO-MOEA/DRVEAKEOO-RVEA
CyberShake301.275 × 10 1 2.333 × 10 1 1.568 × 10 1 2.391 × 10 1 1.113 × 10 1 2.381 × 10 1 1.734 × 10 1 2.814 × 10 1
(1.03 × 10 0 )(3.95 × 10 1 )(9.12 × 10 1 )(5.45 × 10 1 )(1.02 × 10 0 )(2.61 × 10 1 )(1.04 × 10 0 )(6.76 × 10 1 )
503.719 × 10 1 6.274 × 10 1 3.846 × 10 1 6.453 × 10 1 3.810 × 10 1 6.390 × 10 1 4.077 × 10 1 7.199 × 10 1
(1.33 × 10 0 )(8.70 × 10 1 )(1.64 × 10 0 )(7.01 × 10 1 )(1.40 × 10 0 )(1.01 × 10 0 )(1.20 × 10 0 )(1.76 × 10 0 )
1001.433 × 10 2 2.453 × 10 2 1.695 × 10 2 2.505 × 10 2 1.240 × 10 2 2.479 × 10 2 1.464 × 10 2 2.678 × 10 2
(6.95 × 10 0 )(3.42 × 10 0 )(9.06 × 10 0 )(4.02 × 10 0 )(2.51 × 10 0 )(2.63 × 10 0 )(7.05 × 10 0 )(5.68 × 10 0 )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xing, L.; Wu, R.; Chen, J.; Li, J. Knowledge-Based Evolutionary Optimizing Makespan and Cost for Cloud Workflows. Mathematics 2023, 11, 38. https://doi.org/10.3390/math11010038

AMA Style

Xing L, Wu R, Chen J, Li J. Knowledge-Based Evolutionary Optimizing Makespan and Cost for Cloud Workflows. Mathematics. 2023; 11(1):38. https://doi.org/10.3390/math11010038

Chicago/Turabian Style

Xing, Lining, Rui Wu, Jiaxing Chen, and Jun Li. 2023. "Knowledge-Based Evolutionary Optimizing Makespan and Cost for Cloud Workflows" Mathematics 11, no. 1: 38. https://doi.org/10.3390/math11010038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop