Next Article in Journal
A Probabilistic Version of Eneström–Kakeya Theorem for Certain Random Polynomials
Next Article in Special Issue
Multi-Criteria Analysis for the Evaluation of Urban Freight Logistics Solutions: A Systematic Literature Review
Previous Article in Journal
Deep Reinforcement Learning for the Agile Earth Observation Satellite Scheduling Problem
Previous Article in Special Issue
A Heuristic Model for Spare Parts Stocking Based on Markov Chains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Two-Machine Learning Date Flow-Shop Scheduling Problem with Heuristics and Population-Based GA to Minimize the Makespan

1
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
2
Department of Statistics, Feng Chia University, Taichung 40724, Taiwan
3
Department of Industrial Engineering Automation Operation Intelligence, Micron Memory Taiwan Co., Ltd., Taichung 42152, Taiwan
4
Department of Industrial Engineering and Management, Chin-Yi University of Technology, Taichung 411030, Taiwan
5
Department of Industrial Engineering and Management, Cheng Shiu University, Kaohsiung City 83347, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(19), 4060; https://doi.org/10.3390/math11194060
Submission received: 30 August 2023 / Revised: 16 September 2023 / Accepted: 19 September 2023 / Published: 25 September 2023
(This article belongs to the Special Issue Mathematical Programming, Optimization and Operations Research)

Abstract

:
This paper delves into the scheduling of the two-machine flow-shop problem with step-learning, a scenario in which job processing times decrease if they commence after their learning dates. The objective is to optimize resource allocation and task sequencing to ensure efficient time utilization and timely completion of all jobs, also known as the makespan. The identified problem is established as NP-hard due to its reduction to a single machine for a common learning date. To address this complexity, this paper introduces an initial integer programming model, followed by the development of a branch-and-bound algorithm augmented with two lemmas and a lower bound to attain an exact optimal solution. Additionally, this paper proposes four straightforward heuristics inspired by the Johnson rule, along with their enhanced counterparts. Furthermore, a population-based genetic algorithm is formulated to offer approximate solutions. The performance of all proposed methods is rigorously evaluated through numerical experimental studies.

1. Introduction

Learning rates are used in scheduling problems because they are a key parameter in training machine learning models, including those used in optimization problems such as scheduling. In scheduling problems, the goal is to assign tasks or activities to resources in a way that minimizes some objective function, such as the makespan (the length of time it takes to complete all tasks) or the total cost of using the resources. Machine learning algorithms can be used to optimize scheduling decisions by learning from historical data or by simulating different scheduling scenarios (Bottou [1]; Goodfellow et al. [2]; Pinedo [3]).
Learning rates are a critical parameter in the training process of machine learning algorithms because they control the rate at which the model adapts to new information. A high learning rate will cause the model to make large adjustments to the weights or parameters of the model based on each new data point, which can lead to overfitting and poor generalization performance. On the other hand, a low learning rate will cause the model to make very small adjustments based on each data point, which can lead to slow convergence and may cause the model to become stuck in local minima. Therefore, selecting an appropriate learning rate is critical for achieving good performance in scheduling problems and other optimization tasks. The learning rate must be tuned carefully to balance the competing objectives of fast convergence and good generalization, and different learning rate schedules may be used to adapt the learning rate over time as the optimization process progresses (Ruder [4]; Russell and Norvig [5]).
The integration of the learning concept into scheduling research was first introduced by Biskup [6] and Cheng and Wang [7], marking significant milestones in this area of study. Since then, there has been extensive research on scheduling problems that incorporate various types of learning effects (Kuo and Yang [8], Koulamas and Kyparisis [9], Wang [10], Yin et al. [11], Wu et al. [12,13,14], Wang and Wang [15], Niu et al. [16], Zhang et al. [17], etc.). To explore this topic further, readers may refer to the works of Biskup [18], Janiak et al. [19], and Azzouz et al. [20] for a comprehensive survey on incorporating learning effects into scheduling. These resources provide detailed insights into the progress made in this field and offer valuable information for those interested in understanding the impact of learning on scheduling processes.
In the realm of minimizing the makespan on a two-machine flow-shop with learning considerations, various researchers have made significant contributions. Wang and Xia [21] introduced flow-shop scheduling problems with position-based learning effects, aiming to minimize either the makespan or total flowtime. They proposed a heuristic algorithm with a worst-case bound, denoted as m, where m represents the number of machines. Additionally, they developed polynomial algorithms to address specific cases. Expanding on this line of study, Xu et al. [22] investigated flow-shop scheduling problems with position-based learning effects, focusing on minimizing three regular performance criteria: total weighted completion time, discounted total weighted completion time, and the sum of quadratic job completion times. They presented algorithms tailored to their proposed problems and analyzed the worst-case bounds of these algorithms.
Addressing the makespan two-machine problem with a truncated sum of processing-times-based learning function, Wu et al. [23] devised both a branch-and-bound algorithm and a genetic heuristic-based algorithm. Shifting gears to flow-shop scheduling problems with exponential learning effects, Wang and Wang [24] sought to minimize four regular performance criteria: total completion time; total weighted completion time; discounted total weighted completion time; and the sum of quadratic job completion times. They developed heuristic algorithms utilizing optimal permutations derived from corresponding single-machine scheduling problems and discussed the worst-case bounds of each algorithm. In another study, Cheng et al. [25] proposed a branch-and-bound algorithm along with three crossover-based genetic algorithms to tackle a two-machine flow-shop scheduling problem with a truncated learning function.
Furthermore, Wang et al. [26] presented heuristic algorithms and analyzed their worst-case bounds for minimizing various criteria in flow-shop scheduling with truncated position-based learning effects. These criteria included total completion time, makespan, total weighted completion time, discounted total weighted completion time, the sum of quadratic job completion times, and maximum lateness. In a more recent work, Wang et al. [27] applied several heuristics and a branch-and-bound algorithm to address permutation flow-shop problems with a truncated exponential sum of logarithm processing times based and position-based learning effects. They focused on minimizing the makespan and total weighted completion time, respectively.
For some more recent and relevant instances of the two-machine flop-shop scheduling problem with different assumptions or constraints, researchers have proposed various approaches. Hsu et al. [28] addressed a robust version of the two-machine flow-shop scheduling problem, considering scenario-dependent processing times to minimize the total completion time. They introduced a branch-and-bound method, along with several heuristics and their corresponding counterparts. Additionally, they explored several variants of a cloud theory-based simulated annealing algorithm to tackle the problem. Taking into account time-of-use tariff assumptions, Ho et al. [29] developed three heuristics based on Johnson’s rule and three heuristics based on Hadda’s algorithm. Their aim was to jointly optimize the makespan and electricity cost in the context of a two-machine flow-shop scheduling problem under varying electricity pricing. Addressing the factor of resource recycling, Lo and Lin [30] delved into the relocation problem within a two-machine flow-shop scheduling scenario. They introduced a second machine dedicated to recycling resources returned by jobs completed on the first machine. To tackle this, they applied an integer programming model and employed ant colony optimization techniques. Furthermore, Chen et al. [31] proposed a dynamic programming approach and discussed the worst-case performance ratio for a two-machine flow-shop aiming to maximize the total early work subject to a common due date. To address convex resources to the model, Choi and Park [32] discussed and analyzed the approximation to minimize the sum of the makespan and the total resource consumption cost. The summary of these papers is recorded in Table 1.
These studies demonstrate the breadth of research efforts dedicated to understanding and optimizing flow-shop scheduling problems with learning considerations, emphasizing the development of heuristic algorithms, polynomial algorithms, branch-and-bound approaches, and worst-case bound analyses. Several contributions of this paper are provided as follow:
  • We introduce the two-machine flow-shop scheduling problem with step-learning, where job processing times decrease if they start after their job-dependent learning dates;
  • We establish that the proposed scheduling problem is NP-hard, which implies that finding an exact optimal solution is computationally challenging;
  • We present an integer programming model as a method to address the scheduling problem, providing a theoretical framework for optimization.
We devise a branch-and-bound algorithm incorporating two lemmas and a lower bound to facilitate finding an exact optimal solution to the problem and propose four simple heuristics based on the ideas of Johnson’s rule, offering practical, faster solutions. We also introduce improved versions of the heuristics, likely enhancing their performance. Additionally, we develop a genetic algorithm to provide approximate solutions to the scheduling problem, demonstrating a metaheuristic approach for tackling the complexity of the problem.
The remainder of this article is structured as follows: Section 2 introduces the necessary notations and presents the problem that is being investigated. In Section 3, an integer programming formulation is proposed for the model, accompanied by the derivation of two lemmas and a lower bound. These components are integral to the branch-and-bound method employed for finding an optimal solution. Moving on to Section 4, we present four modified heuristics inspired by the Johnson rule, as well as a genetic algorithm (GA) designed to find approximate solutions. Section 5 focuses on the exploration of the proposed GA’s parameters. To assess the effectiveness and efficiency of all the methods discussed, Section 6 presents a series of experimental tests. Finally, the last section offers conclusive remarks and provides insightful suggestions.

2. Notations and Problem Definition

First, several symbols to be used in this study are defined as follows:
  • M = {M1, M2}: represents a collection of two machine codes;
  • N = {J1, J2, , Jn}: represents a collection of n labeled jobs;
  • P 1 j and P 2 j : represent the actual processing time of job j on machine 1 and machine 2, respectively;
  • L D 1 j and L D 2 j : represent the learning dates for job j on machine 1 and machine 2, respectively;
  • u 1 j and u 2 j : represent the original processing times of job Jj on machine 1 and machine 2 in σ ;
  • v 1 j and v 2 j : represent the processing time of job Jj on machine one and machine two, respectively, under the learned schedule via σ for the jobs sorted by order;
  • σ , σ : represent two full schedules of jobs;
  • π 1 , π 2 : represent two partial subsequences of jobs;
  • s 1 and s 2 : represent the starting times for a job on machine 1 and machine 2 in σ ;
  • C k i ( σ ) and C k j ( σ ) : represent the completion times of job i and job j on machine k in schedule σ , k = 1, 2;
  • C k j ( σ ) and C k i ( σ ) : represent the completion times of job i and job j on machine k in schedule σ , k = 1, 2;
  • [ ]: represents the position of the job in a certain schedule.
    C m a x ( σ ) = m a x C 2 [ 1 ] σ , C 2 [ 2 ] σ , , C 2 [ n ] ( σ ) ;
The problem under investigation pertains to the scheduling of two flowline machines, with n jobs waiting to be processed online without any preparation time. Once a job starts processing, it must be completed before the next job can be processed, with no interruptions allowed during the processing. The processing mode for each job is that it must be processed on machine 1 before it can be processed on machine 2. Let P 1 j and P 2 j denote the processing times of job j on machine 1 and machine 2, respectively. The study also considers the types of learning effects, including temporal dependencies and job-dependent learning effects, referred to as job-dependent learning schedules. Specifically, the original processing time for job j on machine 1 is denoted by u 1 j , while v 1 j represents the reduced processing time due to the learning date, such that the processing time required under the learning date ( v 1 j ) is less than or equal to the original processing time ( u 1 j ) or ( v 1 j u 1 j ). Similarly, u 2 j and v 2 j represent the original processing time and processing time required under the learning effect for job j on machine 2, respectively, with v 2 j u 2 j . This study assumes that L D 1 j and L D 2 j are the learning dates for job j on machine 1 and machine 2, respectively. The definition of the actual processing time of job j on machine 1 and machine 2 can be given by synthesizing the above explanations as follows:
P 1 j = { u 1 j , s 1 < L D 1 j v 1 j , s 1 L D 1 j   and   P 2 j = u 2 j , s 2 < L D 2 j v 2 j , s 2 L D 2 j ,
where s 1 and s 2 represent the starting times of job j on machine 1 and machine 2. The objective function of this study is to find a schedule to minimize the makespan of n given jobs. The problem can be simplified as F2 / / v 1 j u 1 j ,   v 2 j u 2 j L D 1 j ;   L D 2 j / / C m a x .

3. Exact Solution Methods

In this section, we consider using mixed integer programming to mathematically model the proposed problem and utilize CPLEX software to obtain the optimal solution. The initial modeling approach is described as follows:
For all possible ordering problems of n jobs, there are L = n! different arrangements. Considering n jobs, each of which is processed on two machines in a certain order, the same ordering is used for processing on both machines.
f σ * = m i n { f σ 1 , f σ 2 , , f σ J , , f σ L }
In Equation (1), f σ J represents the objective function, which is the completion time under the job sequencing plan J. The goal of this project is to find the job sequencing plan σ * that minimizes the objective function value among all possible plans, as well as the corresponding optimal objective function value f σ * .
f σ J = X 1 k p k + i = 1 n k = 1 n X i k ( t i k + q i k )
In Equation (2), when the job numbered k undergoes processing at the ith position, the following variables are defined:
X i k is a 0–1 variable used to indicate whether the job k is processed at the ith position.
X i k = 1 ,   the   job   k   is   processed   at   the   i th   position   0 ,   o t h e r w i s e   ,   i , k = 1,2 , , n
i = 1 n X i k = 1 , 2 , , n ,
Equation (3) ensures that only one task is chosen to be executed at each time instant.
k = 1 n X i k = 1 , 2 , , n ,
Equation (4) ensures that each task is executed only at one time instant.
p i k denotes the processing time of job k on machine 1.
p i k = u 1 i , A i k < L D 1 i v 1 i , A i k L D 1 i
q i k denotes the processing time of job k on machine 2.
q i k = u 2 i , B i k < L D 2 i v 2 i , B i k L D 2 i
A i k denotes the starting time of job k placed on ith position on machine 1.
A i k =   0   ,   i = 1   j = 1 i 1 m = 1 n X j m p j m , 2 i n
B i k denotes the starting time of job k placed on ith position on machine 2.
B i k = A i k + k = 1 n X i k p i k + t i k
C i k denotes the completion time of job k to be processed on the ith position on machine 2.
C i k = B i k + k = 1 n X i k q i k
t i k represents the waiting time required for the job k after completing processing in machine 1 and before starting processing in machine 2.
t i k = 0 ,   i = 1     Y i k ( C ( i 1 ) l A i k k = 1 n X i k p i k ) ,   i = 2 ,   , n
Y i k is a binary variable (0 or 1) used to determine whether the job k needs to wait when being processed in the ith sequence in machine 2.
Y i k = 0 , C i 1 l < A i k + k = 1 n X i k p i k   a n d   2 i n 1 , C i 1 l A i k + k = 1 n X i k p i k   a n d   2 i n
l 1,2 , , n   and   l k

Branch-and-Bound Method

In what follows, we will propose several dominances based on the characteristics of the completion time of the workpieces above to cooperate with the branch-and-bound method to find the optimal solution.
For problem F2 / / v 1 j u 1 j , v 2 j u 2 j L D 1 j ; L D 2 j / / C m a x : Assuming two schedules σ 1 = π 1 , J i , J j , π 2 and σ 2 = π 1 , J j , J i , π 2 , and assuming that J L is the last job in partial schedule π 1 , s 1 and s 2 are the starting times of job J L on machine 1 and machine2. Regarding this problem, from the above data examples, we preliminarily propose the following characteristics: To show that σ 1 dominates σ 2 , one suffices to show that C 1 j ( σ 1 ) C 1 i ( σ 2 ) , and C 2 j σ 1 C 2 i ( σ 2 ) .
Lemma 1. 
If the jobs  J i  and  J j   are present in both  σ 1 and  σ 2 , s 1 + u 1 j m i n { L D 1 i ,   L D 1 j } , max s 1 + u 1 j , s 2 + u 2 i < m i n { L D 2 i ,   L D 2 j } , u 1 i u 1 j  and  u 2 j u 2 i , then  σ 1  dominates  σ 2 .
Proof. 
Since s 1 + u 1 i s 1 + u 1 j m i n { L D 1 i ,   L D 1 j } , it implies s 1 < L D 1 i and s 1 + u 1 j < L D 1 j . Thus, one has C 1 i ( σ 1 ) = s 1 + u 1 i and C 1 j ( σ 1 ) = s 1 + u 1 i + u 1 j .
On the other hand, since s 1 + u 1 j m i n { L D 1 i , L D 1 j } , it implies s 1 < L D 1 j and s 1 + u 1 j < L D 1 j . Thus, one has
C 1 j ( σ 2 ) = s 1 + u 1 j   and   C 1 i ( σ 2 ) = s 1 + u 1 j + u 1 i = C 1 j ( σ 1 ) .
From the given max s 1 + u 1 j , s 2 + u 2 i < m i n { L D 2 i , L D 2 j } , and u 2 j < u 2 i , one has
C 2 i σ 1 = max C 1 i ( σ 1 ) , s 2 = max s 1 + u 1 i , s 2 + u 2 i ,   and   C 2 j σ 1 = max C 1 j σ 1 , C 2 i σ 1 + u 2 j = max s 1 + u 1 i + u 1 j , max s 1 + u 1 i , s 2 + u 2 i + u 2 j = max s 1 + u 1 i + u 1 j + u 2 j , max s 1 + u 1 i , s 2 + u 2 i + u 2 j .
From the given max s 1 + u 1 j , s 2 + u 2 i < m i n { L D 2 i , L D 2 j } , and u 2 j < u 2 i one has
C 2 j σ 2 = max C 1 j σ 2 , s 2 + u 2 j = max s 1 + u 1 j , s 2 + u 2 j ,   and   C 2 i ( σ 2 ) = max C 1 i σ 2 , C 2 j σ 2 + u 2 j = max s 1 + u 1 j + u 1 i , max s 1 + u 1 j , s 2 + u 2 j + u 2 i = max s 1 + u 1 j + u 1 i + u 2 i , max s 1 + u 1 j , s 2 + u 2 j + u 2 i
From the given u 1 i u 1 j and u 2 j u 2 i , comparing the two max terms of C 2 j ( σ 1 ) in Equation (5) with the two max terms of C 2 i ( σ 2 ) in Equation (6), one obtains C 2 j σ 1 C 2 i ( σ 2 ) . □
σ 1 = π 1 , J i , J j , π 2   and   σ 2 = π 1 , J j , J i , π 2 ,
Lemma 2. 
If the jobs  J i  and  J j  are present in both σ 1  and  σ 2 ,  v 1 i v 1 j ,  v 2 j v 2 i ,  m a x { L D 1 i ,   L D 1 j } s 1 ,  s 2 s 1 + v 1 i , and  m i n { s 1 + v 1 i + v 1 j , s 1 + v 1 i + v 2 j , s 1 + v 1 i ,   s 2 + v 2 j } m a x { L D 2 i ,   L D 2 j } , then  σ 1  dominates  σ 2 .
Proof. 
To show that  σ 1  dominates  σ 2 , one suffices to show that  C 1 j ( σ 1 ) C 1 i ( σ 2 ) , and   C 2 j σ 1 C 2 i ( σ 2 ) .
For the schedule  σ 1 , from the given  s 1 L D 1 i , then  C 1 i ( σ 1 ) = s 1 + v 1 i , and from the  L D 1 j s 1 < s 1 + v 1 i , one has  C 1 j ( σ 1 ) = s 1 + v 1 i + v 1 j .
From the given  L D 1 j s 1 + v 1 i , and  s 2 s 1 + v 1 i , one has  C 2 i ( σ 1 ) = max s 1 + v 1 i , s 2 + v 2 i = s 1 + v 1 i + v 2 i .
Owing to  C 1 j ( σ 1 ) = s 1 + v 1 i + v 1 j L D 2 j , and  C 2 i σ 1 = s 1 + v 1 i + v 2 i s 2 + v 2 i s 2 + v 2 j L D 2 j , one then has  C 2 j ( σ 1 ) = max C 1 j σ 1 , C 2 i σ 1 + v 2 j = max s 1 + v 1 i + v 1 j , s 1 + v 1 i + v 2 i + v 2 j .
Thus, one has
C 2 j ( σ 1 ) = max s 1 + v 1 i + v 1 j + v 2 j , s 1 + v 1 i + v 2 i + v 2 j .
For the schedule σ 2 , from the given s 1 L D 1 j , then C 1 j ( σ 2 ) = s 1 + v 1 j , and from the L D 1 i s 1 < s 1 + v 1 j , one has C 1 i ( σ 2 ) = s 1 + v 1 j + v 1 i . Now, one obtains C 1 j ( σ 1 ) = s 1 + v 1 i + v 1 j = C 1 i ( σ 2 ) .
From the given v 1 i v 1 j , then s 2 s 1 + v 1 i s 1 + v 1 j , and L D 2 j s 1 + v 1 i s 1 + v 1 j , one has C 2 j ( σ 2 ) = max s 1 + v 1 j , s 2 + v 2 j = s 1 + v 1 j + v 2 j .
Owing to C 1 i ( σ 2 ) = s 1 + v 1 j + v 1 i L D 2 i , and C 2 j σ 2 = s 1 + v 1 j + v 2 j s 1 + v 1 i + v 2 j L D 2 j , one then has C 2 i ( σ 2 ) = max C 1 i σ 2 , C 2 j σ 2 + v 2 i ; that is
C 2 i ( σ 2 ) = max s 1 + v 1 j + v 1 i + v 2 i ,   s 1 + v 1 j + v 2 j + v 2 i .
From the given v 1 i v 1 j and v 2 j v 2 i , comparing two max terms of C 2 j σ 1 in Equation (7) and two max terms of C 2 i ( σ 2 ) in Equation (8), one obtains C 2 j σ 1 C 2 i ( σ 2 ) . □
In branch-and-bound algorithms, the lower bound plays a crucial role in guiding the search process and improving efficiency. Its primary function is to provide an estimate of the minimum possible value for the objective function or cost in an optimization problem. By using a lower bound, the branch-and-bound algorithm intelligently prunes unpromising branches, reducing the overall search space and improving the chances of finding the optimal solution more efficiently. It acts as a guiding factor to focus the search on the most promising areas, saving computational resources and speeding up the optimization process. Considering these observations, we also propose the following lower bound. Moreover, let σ = π , π c , where π and π c denote the sequence with i scheduled jobs and the sequence with ( n i + 1 ) unscheduled jobs, respectively, and s 1 and s 2 denote the completion times of the last job on machine 1 and 2, respectively, in π . According to the definition of completion time of a job to be scheduled on the (i + 1)-th position on machine 1 and machine 2, we have the following equalities.
C 1 i + 1 ( σ ) = C 1 [ i ] ( σ ) + X [ i + 1 ] u 1 [ i + 1 ] + ( 1 X [ i + 1 ] ) v 1 [ i + 1 ] = s 1 + X i + 1 ( u 1 i + 1 v 1 i + 1 ) + v 1 [ i + 1 ] s 1 + v 1 [ i + 1 ] , C 2 i + 1 ( σ ) = max C 1 i + 1 ( σ )   ,   C 2 i ( σ ) + Y i + 1 u 2 i + 1 + 1 Y i + 1 v 2 i + 1 = max C 1 i + 1 ( σ ) , s 2 + Y i + 1 ( u 2 i + 1 v 2 i + 1 ) + v 2 [ i + 1 ] max s 1 + v 1 [ i + 1 ] ,   s 2 + Y i + 1 ( u 2 i + 1 v 2 i + 1 ) + v 2 [ i + 1 ] s 2 + v 2 [ i + 1 ] .
It is noted that X [ q ] s and Y [ q ] s denote two indicator variables to indicate whether the starting times of a job j assigned to the q-th position is more than the learning dates of the job j on machines 1 and 2, respectively.
L B 1 = s 2 + l = i + 1 n v 2 [ l ] .
In a similar way, another lower bound can be obtained as follows:
C 1 [ i + 1 ] = C 1 [ i ] + X i + 1 ( u 1 i + 1 v 1 i + 1 ) + v 1 [ i + 1 ] = s 1 + X i + 1 ( u 1 i + 1 v 1 i + 1 ) + v 1 [ i + 1 ] s 1 + v 1 [ i + 1 ] , C 2 i + 1 = max C 1 i + 1   ,   C 2 i + Y i + 1 u 2 i + 1 + 1 Y i + 1 v 2 i + 1 max s 1 + v 1 [ i + 1 ]   , s 2 + Y i + 1 ( u 2 i + 1 v 2 i + 1 ) + v 2 [ i + 1 ] s 1 + v 1 [ i + 1 ] + v 2 [ i + 1 ] , C 1 n = s 1 + l = i + 1 n X l u 1 l + l = i + 1 n 1 X l v 1 l s 1 + l = i + 1 n v 1 [ l ] , C 2 [ n ] s 1 + l = i + 1 n v 1 l + v 2 l .
Based on the above Equation (11), we have
L B 2 = s 1 + l = i + 1 n v 1 [ l ] + m a x j π c { v 2 j } .
From Equations (9) and (12), a tighter lower bound (LB) can obtained as follows:
LB = max { s 2 + l = i + 1 n v 2 [ l ] ,     s 1 + l = i + 1 n v 1 [ l ] + m a x j π c { v 2 j } }

4. Heuristics and Population-Based Genetic Algorithm

A heuristic method is a problem-solving strategy that relies on practical and experience-based approaches to find a solution, instead of following a strict algorithmic process. There are several compelling reasons for using a heuristic method to solve complex problems with limited resources. For instance, heuristic methods can offer faster solutions than rigorous methods, and they are particularly useful when facing difficult problems that may not have straightforward solutions.
The Johnson rule is a heuristic method used to solve a two-machine flow-shop scheduling problem with the objective of minimizing the makespan (i.e., F2 / / C m a x ). The makespan is the total time it takes to complete all jobs on both machines. This process continues until all jobs have been scheduled. The Johnson rule guarantees that the resulting schedule is optimal for minimizing the makespan in a two-machine flow-shop scheduling problem. However, the Johnson rule, proposed by Johnson [33], cannot find an optimal solution for the proposed problem due to the fact that the proposed problem F2 / / v 1 j u 1 j , v 2 j u 2 j L D 1 j ; L D 2 j / / C m a x is an NP-hard one. The following are the steps involved in applying Johnson’s rule: Step 1: Identify the set of jobs that need to be processed; Step 2: Determine the smallest processing times across all n jobs and two machines; Step 3: If the smallest processing time from the first machine is at the beginning of the process, schedule that job first. On the other hand, if the smallest processing time from the second machine is at the beginning of the process, schedule that job last; Step 4: Remove the job that has been scheduled from the set of jobs; and Step 5: Repeat steps 1–4 until all jobs have been scheduled. Following Johnson’s rule, we apply the Johnson rule to each of the four different initial inputs: { L D 1 i ,   L D 2 i , i Ω = { 1 , 2 , , n } ; { u 1 i ,   u 2 i , i Ω = { 1 , 2 , , n } } ; { v 1 i ,   v 2 i , i Ω = { 1 , 2 , , n } } ; and { u 1 i ,   v 2 i , i Ω = { 1 , 2 , , n } } and then refine them further using a pairwise interchange method. The complexity of Johnson’s Rule is O(n log n), where n represents the number of jobs, while the complexity associated with the pairwise interchange method is O(n2). Therefore, the computational time of each of four proposed heuristics is O(n2). The details of four proposed heuristics are introduced below:
  • Heuristic 1 (named as J2LD-pi)
  • Step 1: Input a set of { L D 1 i ,   L D 2 i , i Ω = { 1 , 2 , , n } } ;
  • Step 2: Set P 1 i = L D 1 i and P 2 i = L D 2 i , i Ω ;
  • Step 3: Set σ = ( , , , ) , f = 1 , and l = n ;
  • Step 4: Find P 1 i * = m i n { P 1 i , i Ω } and P 2 i * * = m i n P 2 i ,   i Ω ;
  • Step 5: If P 1 i * P 2 i * * , then job i * is assigned to be fth position in a schedule,
  • (i.e., σ = ( i * , , , ) ), f = f + 1, and delete job i * from Ω . Go to Step 7;
  • Step 6: If P 1 i * > P 2 i * * , then job i * * is assigned to be lth position in a schedule,
  • (i.e., σ = ( , , ,   i * * ) ), l = l − 1, and delete job i * * from Ω ;
  • Step 7: If Ω is not empty, return to do Step 4 to Step 6; otherwise, output the final complete schedule σ ;
  • Step 8: Refine the final complete schedule σ by using a pairwise interchange method.
  • Heuristic 2 (named as J2U-pi)
  • Step 1: Input a set of { u 1 i ,   u 2 i , i Ω = { 1 , 2 , , n } } ;
  • Step 2: Set P 1 i = u 1 i and P 2 i = u 2 i , i Ω ;
  • Step 3 to Step 8 are the same as those of J2LD.
  • Heuristic 3 (named as J2V-pi)
  • Step 1: Input a set of { v 1 i ,   v 2 i , i Ω = { 1 , 2 , , n } } ;
  • Step 2: Set P 1 i = v 1 i and P 2 i = v 2 i , i Ω ;
  • Step 3 to Step 8 are the same as those of J2LD.
  • Heuristic 4 (named as J2UV-pi)
  • Step 1: Input a set of { u 1 i ,   v 2 i , i Ω = { 1 , 2 , , n } } ;
  • Step 2: Set P 1 i = u 1 i and P 2 i = v 2 i , i Ω ;
  • Step 3 to Step 8 are the same as those of J2LD.

Genetic Algorithm

Numerous literature studies have consistently shown that heuristics, in general, are characterized by their relative simplicity and ease of construction. On the other hand, metaheuristics are notably more complex and challenging to design and employ effectively, particularly when it comes to intelligent random search strategies (Holland [34]; Essafi et al. [35]; Nguyen et al. [36]; Fan et al. [37]; Tutumlu and Saraç [38]; Wang and Chen [39]). Among the various metaheuristics, the genetic algorithm (GA) stands out as a widely applied and successful approach for obtaining high-quality approximate solutions to a diverse range of combinatorial problems.
The GA capitalizes on genetic and neural selection mechanisms, including mutation, crossover, and reproduction, to conduct an efficient search for optimal solutions in complex problem domains. The effectiveness of these genetic operators has been particularly evident in solving numerous NP-hard combinatorial problems. By leveraging these operators, the GA iteratively refines and explores the solution space, gradually converging towards high-quality solutions. The GA’s ability to strike a balance between exploration and exploitation enables it to effectively navigate large solution spaces and uncover near-optimal solutions. This makes it a valuable tool for addressing complex optimization problems where traditional approaches may struggle.
In the GA framework, we utilize a set of continuous real numbers to represent order codes through random number encoding. For instance, a chromosome (0.63, 0.51, 0.21, 0.53, 0.18) can be decoded using the ranking method into a schedule (5, 3, 2, 4, 1). During the reproduction stage, parents are selected, and a specific crossover operator, such as linear order crossover (Iyer and Saxena [40]; Wu et al. [41]), is employed to generate offspring. Alternatively, readers can refer to Larranaga et al. [42] for a partially mapped crossover operator or Kellegoz et al. [43] or Castelli et al. [44] for a single crossover point operator. To introduce diversity, we incorporate the fourth displacement mutation (referred to as M4) proposed by Nearchou [45]. This mutation involves randomly selecting a subsequence of orders and inserting them at a random position. In the context of our study, let n_size represent the number of parents, Pm denote the mutation rate, and g_size represent the number of iterations or generations for executing the GA. Now, let us summarize the main structures of the proposed GA:
  • The details of GA:
  • Step 1. Initialize parameters: Set the values of n_size (number of parents), Pm (mutation rate), and g_size (number of iterations);
  • Step 2. Generate an initial population of n_size parents (schedules) and evaluate their fitness values;
  • Step 3. Iterate i from 1 to g_size:
-
Select two parents from the population using the roulette wheel method;
-
Apply linear order crossover to produce a set of n_size offspring;
-
For each offspring, generate a random number r (0 < r < 1). If r is less than Pm, perform a displacement mutation to create a new offspring;
-
Keep track of the best schedule found so far;
-
Replace the current population of parents with their offspring;
  • Step 4. End the iteration loop (g_size);
  • Step 5. Output the final best schedule and its corresponding fitness value.
By following these steps, the GA gradually evolves a population of schedules over multiple iterations, utilizing crossover and mutation operators to explore the search space and improve the quality of the solutions. The algorithm terminates after reaching the specified number of iterations, providing the best schedule found and its associated fitness value as the output.

5. The Parameter Exploration for GA

In this section, two different values of n are utilized for parameter exploration of GA: a large n = 400 and a small n = 10. Additionally, the learning dates control parameters for both machines, denoted as w1 and w2, are set to 0.5. For the scenarios with large n = 400 and small n = 10, one hundred instances are generated for each case. Data analysis is conducted for each dataset, involving the processing times u 1 i and u 2 i of jobs on machine M1 and machine M2, respectively. These times are uniformly distributed within the range U(51, 100). Under the influence of learning effect, the processing times v 1 i and v 2 i for jobs on machine M1 and machine M2 are generated from U(1, 50). The learning dates L D 1 and L D 2 for machine M1 and machine M2 are derived from U(1, U(1, w 1 i = 1 n u 1 i )) and U(1, w 2 i = 1 n u 2 i ), respectively.
In this data context, we compute both the optimal solutions (Oi) obtained through the branch-and-bound method and the approximate solutions (Hi) obtained using the genetic algorithm (GA). The average error percentage (AEP) is calculated as A AEP = { 1 100 H i O i / O i } × 100 % to investigate the simulated parameter combinations (such as evolution generations Gsize, mutation rate p, and chromosome count n_size) within the GA algorithm. The experimental approach entails initially fixing two parameters and then varying the third parameter as the experimental subject for testing and exploration. This process continues until the minimum AEP value is identified. The parameter is then treated as a fixed parameter, and the aforementioned method is applied to identify the next suitable parameter. This iterative process persists until appropriate values are determined for all parameters.
First, the scenario with n = 10 was tested. Figure 1a depicts the parameter exploration of evolution generations (g_size) under the condition of a fixed mutation rate of p = 0.02 and a chromosome count of n_size = 500. The g_size was varied from 20 to 360, with increments of 20 generations. From Figure 1a, it is evident that the lowest point of the average error percentage (AEP) was reached at 240 generations. Therefore, 240 generations was selected as the value for the g_size for subsequent simulation testing.
Based on the above findings, the optimal number of evolution generations was determined to be 240. Therefore, with the evolution generations fixed at g_size = 240 and the mutation rate at p = 0.02, we conducted an exploration of the chromosome count parameter (n_size). Specifically, we varied the chromosome count (n_size) from a minimum of 10 chromosomes to a maximum of 200 chromosomes, increasing by 10 chromosomes in each step. In Figure 1b, it can be observed that the GA reached a relatively stable state around 100 chromosomes. However, the lowest point occurred at 180 chromosomes. Hence, we selected 180 chromosomes as the reference value for the n_size in the subsequent GA simulation testing.
Based on the results above, we determined that suitable values were g_size = 240 for evolution generations and n_size = 180 for chromosome count. Therefore, in the subsequent steps, we fixed g_size = 240 and n_size = 180 to explore the parameter range of mutation rate (p). We considered a mutation rate range from 0.01 to 0.2, with increments of 0.01. As evident from Figure 1c, it became evident that the GA achieved the minimum error at a mutation rate of 0.16. Consequently, this study employed a mutation rate of 0.16 as the basis for GA testing in the subsequent phases.
Next, the initial setup for parameter exploration involving a larger n was conducted with n = 400. The branch-and-bound method was utilized to calculate the root lower bound value (LBi), and the genetic algorithm (GA) was used to compute approximate solutions (Hi). The percentage gap in their differences (average gap percentage or GAP) was computed as GAP = 1 100 H i L B i / L B i × 100 % to explore the parameter values of g_size (evolution generations), p (mutation rate), and n_size (chromosome count) for the GA. The goal was to determine the optimal values of g_size, p, and n_size that yielded the most suitable GAP values. These optimal values were intended for reference in subsequent GA algorithm simulations for larger n.
As shown in Figure 1d, when n = 400, with n_size = 2000 fixed and a mutation rate of p = 0.02, the exploration of the evolution generations (g_size) ranges from a minimum of 100 generations to a maximum of 2000 generations, increasing by 100 generations in each step. From the graph, it is evident that the lowest GAP value occurred at around 1500 generations. Consequently, the evolution generations for the subsequent GA simulations was set to g_size = 1500.
Subsequently, the exploration of the mutation rate was conducted with n = 400, g_size = 1500, and n_size = 2000 held constant. The range of mutation rates for investigation was set from 0.01 to 0.2, increasing by increments of 0.01. From Figure 1e, it becomes evident that the mutation rate p reached its lowest GAP value at 0.16. Therefore, the subsequent GA algorithm adopted a mutation rate setting of p = 0.16.
Finally, with n = 400, g_size = 1500, and p = 0.16 fixed, we explored the parameter n_size (chromosome count). Figure 1f illustrates that when setting the chromosome count from 400 to 4000, with increments of 400 chromosomes in each step for simulation testing, the results indicated a progressively increasing trend in GAP values. Consequently, we considered an exploration range for an n_size between 100 and 400 chromosomes, increasing by 100 chromosomes in each step. Additionally, we also assessed the performance of GA with a chromosome count of 50. Observing the graph, it is apparent that the lowest GAP value occurred at a chromosome count of 200. Therefore, the decision was taken to utilize 200 chromosomes as the parameter for subsequent GA testing.

6. Simulation Results for Proposed Algorithms

In this section, we conducted several experimental designs to evaluate the performances of the four proposed algorithms, a population-based genetic algorithm, and a branch-and-bound method. The designs included six different values of n with n = 8, 10, and 12 for the small number of jobs and n = 100, 200, and 400 for the big number of jobs, along with nine cases of learning date control parameters (w1, w2) with (0.25, 0.25), (0.25, 0.5), (0.25,0.75), (0.5, 0.25), (0.5, 0.5), (0.5,0.75), (0.75, 0.25), (0.75, 0.5), and (0.75,0.75). For each case, the processing times u 1 i and u 2 i of jobs on machine M1 and machine M2 were generated from within the range U(51, 100), uniformly distributed, respectively. Meanwhile, under the influence of learning effect, the processing times v 1 i and v 2 i for jobs on machine M1 and machine M2 were generated from U(1, 50), respectively. Additionally, the learning dates L D 1 and L D 2 for machine M1 and machine M2 were derived from U(1, U(1, w 1 i = 1 n u 1 i )) and U(1, w 2 i = 1 n u 2 i ), respectively. This study reports 54 cases and one hundred instances for each case. In total, we generated 5400 tested instance problems where 2700 instances were for the small number of jobs and for the big number of jobs, respectively. The results of six different values of n with n = 8, 10, and 12 for the small number of jobs and n = 100, 200, and 400 for the big number of jobs are reported in Table 2 and Table 3.
For the cases of the small number of jobs, we computed both the optimal solutions (Oi) obtained through the branch-and-bound method (B&B) and the approximate solutions (Hi) obtained using the four heuristics (J2LD-pi, J2U-pi, J2V-pi, J2UV-pi) and genetic algorithm (GA). The average error percentage (AEP) was calculated as AEP= { 1 100 H i O i / O i } × 100 % and standard deviation (STD) of errors to determine the performances of all the proposed algorithms over the different values of n, w1, and w2. Regarding the performance of B&B, we reported the mean and maximum of explored nodes and their corresponding CPU time (in second) for each case. Additionally, we designed a tested instance problem as not feasible one when the B&B consumed the number of nodes up to 10 9 . The effects of n, w1, and w2 in Table 4 and Table 5 were obtained from Table 2.
For the performance of B&B, Table 4 shows that B&B took more nodes at w 2 = 0.75 than those at w 2 = 0.25 or 0.5. The pattern of w 1 was similar to that of w 2 , except n = 8. The pattern became clear as n increases. It also can be observed in Table 2 that B&B consumed fewer nodes at w 1 = 0.5 or w 2 = 0.5 . We also noted that in the two cases of n = 10, with different weight configurations w   1 = 0.25 and w 1 = 0.5, although the total mean node count was higher in the case of w 1 = 0.25, this is because in this case, the maximum node count for the weight configuration ( w = 0.5, w 2 = 0.75) was even greater than the maximum node count for ( w 1 = 0.25, w 2 = 0.75). However, the time taken to solve these two scenarios was remarkably similar. This is why, in the scenario with w 1 = 0.25, despite having a higher total mean node count, the time required to solve the problem was less. Overall, the nodes grew very large as n increased. It can be confirmed this problem is an NP-hard one.
With respect to the achieved results of the four proposed heuristics and the genetic algorithm (GA) applied to the case of small n, Table 5 and Figure 2 illustrate their AEPs. In summary, the mean AEP values were 7.23, 7.13, 6.23, 5.64, and 0.25 for J2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA, respectively. It is evident that the GA algorithm outperformed all of the other four heuristics.
Table 5 further presents the means of AEP values corresponding to different levels of the factors n, w 1 , and w 2 . An observation from Table 5 revealed that the AEPs of the four heuristics and the GA algorithm exhibited a slight increase as n increased from 8 to 12. Conversely, these AEPs remained relatively unaffected by variations in the values of w 1 or w 2 . Turning to the computational times, Figure 3 demonstrates that all four heuristics and the GA algorithm required less than 1 s to execute.
Statistical evidence to determine the significance of differences among the four heuristics and the GA algorithm is discussed below. Initially, an analysis of variance (ANOVA) method using a linear model was employed on the AEP data within SAS 9.4. However, the normality assumption of the linear model was found to be violated. This was evident from the Kolmogorov–Smirnov normality test, which yielded a p-value less than 0.05 (with a statistic value of 0.084). Consequently, due to this non-normality, the data were handled using a nonparametric approach. Specifically, for the 32 instances of the (n*w1*w2) problem, the Friedman test (with p-value less than 0.0001) was conducted based on ranks of AEP values. The results of the test provided evidence that the AEP samples were not drawn from the same distribution at a significance level of α = 0.05.
To further assess 5 2   pairwise differences among the four heuristics and the GA algorithm, the WNMT (Wilcoxon–Nemenyi–McDonald–Thompson) procedure was introduced. Table 6 (column 3) illustrates the sum of ranks of AEP for the 32 instance problems across the four heuristics and the GA algorithm. The rank sums for J2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA were 118, 117, 88, 55, and 27, respectively. Statistically, there was no significant difference among the groups of J2LD_pi, J2U_pi, and J2V_pi at a significance level of 0.05. However, J2UV_pi demonstrated a significant improvement over each of the J2LD_pi, J2U_pi, and J2V_pi heuristics. Additionally, the GA algorithm exhibited significant superiority compared to each of the J2LD_pi, J2U_pi, J2V_pi, and J2UV_pi heuristics. Notably, the GA yielded the smallest AEP and rank sum for a small number of jobs.
Concerning the obtained results in Table 3 for the proposed four heuristics and the GA algorithm in the context of large n, Table 6 (summarized from Table 3) and Figure 4 provide an account of the average Relative Percentage Deviation (RPD) for the four heuristics and the GA algorithm. It is evident from Table 7 that, regardless of the value of n, the GA consistently achieved the lowest average RPD among all job sequences. The average RPD values for J2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA were 8.65, 7.09, 8.33, 6.84, and 0.14, respectively. Furthermore, Figure 4 presents box plots depicting the distribution of average RPD for both the four heuristics and the GA algorithm. These box plots indicate that the interquartile ranges of average RPDs for J2LD_pi, J2U_pi, J2V_pi, and J2UV_pi in comparison to the GA algorithm were broader. It is also noteworthy from Table 3 that as the value of n increased from 100 to 400, the average RPD values of J2LD_pi, J2U_pi, J2V_pi, and J2UV_pi decreased. Conversely, the average RPD of the GA algorithm experienced only a slight increase.
A linear model was applied to the average RPD observations in order to analyze the distinctions between the four heuristics and a GA algorithm. Nevertheless, the assumption of normality for the RPD samples was violated, as indicated by p-values lower than 0.05 in the Kolmogorov–Smirnov normality test. The calculated statistic was 0.096. Consequently, the Friedman’s test with a chi-square value of 69.1532 (p-value of less than 0.0001) was employed to explore the pairwise variations among all proposed algorithms. Examining the sum of ranks of the average RPD computed for each of the 27 blocks related to the test problem instances, these outcomes validated that the average RPD samples did not adhere to the same distribution. In other words, J2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA exhibited distinct search performances.
Table 7 (column 5) displays the rank-sum of average RPDs for J2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA. To conduct a more detailed comparison of ( 5 2 ) pairwise differences among 2LD_pi, J2U_pi, J2V_pi, J2UV_pi, and GA, the WNMT test (with a significance level of 0.05) was employed. The results indicated that GA achieved the best performance, being placed in the group with the lowest rank-sum of 31. On the other hand, the heuristics (J2LD_pi, J2U_pi, J2V_pi, J2UV_pi) were categorized in the group with higher rank-sums (76, 112), implying inferior performance. This test confirmed that, at a significance level of α = 0.05, the GA algorithm statistically outperformed each of the four heuristics individually. Overall, the GA algorithm demonstrated the most effective search capability, particularly for scenarios involving a large number of jobs.
Regarding computational time, Figure 5 displays the box plots of CPU times (in seconds) for the five proposed algorithms. On average, the GA algorithm required a mean CPU time of 10.36 s to solve an instance problem, whereas each of the J2LD_pi, J2U_pi, J2V_pi, and J2UV_pi heuristics necessitated an average CPU time of 0.32 s to solve an instance problem. This discrepancy in CPU time can be attributed to the heightened computational demands of the GA algorithm when compared to each of the J2LD_pi, J2U_pi, J2V_pi, and J2UV_pi heuristics. The underlying factors contributing to the increased CPU time for GA include its elevated complexity, the need for diversity maintenance, adaptation, global search, and problem-specific techniques intrinsic to metaheuristic approaches.

7. Conclusions and Suggestions

This study delved into the scheduling of the two-machine flow-shop problem using step-learning. In this context, job processing times decrease if they commence after their job-dependent learning dates. Initially, we concentrated on a scenario where there was no allowance for idle time between successive jobs on the first machine, yet idle periods were permissible on the second machine. The main goal of this investigation was to optimize resource allocation and task sequencing, aiming to maximize time utilization efficiency and ensure the timely completion of all jobs, a shared objective in the pursuit of scheduling process optimization.
For this challenging problem, we proposed a Mixed-Integer Programming (MIP) model. Subsequently, we employed a branch-and-bound method with two dominance properties and a lower bound to generate optimal solutions for cases with a small number of jobs (e.g., n = 8, 10, and 12). This method performed well up to n = 12 jobs, which encompassed various combinations of job classes and controllable learning indices w1 and w2 for learning dates. For cases with a larger number of jobs, such as n = 100, 200, and 400, we introduced four problem-based heuristics, each accompanied by a pairwise interchange scheme, along with a genetic algorithm (GA) for obtaining approximate solutions. It was observed that the GA outperformed the other four heuristics in both small n and large n tested instances.
One of the issues for future study is to expand the production environment to a two-stage three-machine flow-shop production process with learning dates. Additionally, it is worthwhile to consider extending the same setup to minimize the total completion times of all provided jobs.

Author Contributions

Conceptualization, Y.-H.C.; Methodology, J.-Y.X., W.-C.L., J.-H.C. and C.-C.W.; Software, Y.-W.C.; Investigation, J.-H.C.; Resources, J.-Y.X.; Data curation, Y.-W.C.; Writing—original draft, C.-C.W.; Writing—review & editing, W.-C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China, grant number 72271048, and in part by the Ministry of Science and Technology of Taiwan, NSTC 112-2221-E-035-060-MY2.

Acknowledgments

Authors thank the editor and two referees for their useful and valuable suggestions and comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bottou, L. Stochastic gradient descent tricks. In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 2012; pp. 421–436. [Google Scholar]
  2. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  3. Pinedo, M. Scheduling: Theory, Algorithms, and Systems; Springer Science & Business Media: Berlin, Germany, 2012. [Google Scholar]
  4. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  5. Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach; Prentice Hall: Hoboken, NJ, USA, 2010. [Google Scholar]
  6. Biskup, D. Single-machine scheduling with learning considerations. Eur. J. Oper. Res. 1999, 115, 173–178. [Google Scholar] [CrossRef]
  7. Cheng TC, E.; Wang, G. Single machine scheduling with learning effect considerations. Ann. Oper. Res. 2000, 98, 273–290. [Google Scholar] [CrossRef]
  8. Kuo, W.H.; Yang, D.L. Minimizing the total completion time in a single- machine scheduling problem with a time- dependent learning effect. Eur. J. Oper. Res. 2006, 174, 1184–1190. [Google Scholar] [CrossRef]
  9. Koulamas, C.; Kyparisis, G.J. Single-machine and two-machine flowshop scheduling with general learning functions. Eur. J. Oper. Res. 2007, 178, 402–407. [Google Scholar] [CrossRef]
  10. Wang, J.B. Single-machine scheduling problems with the effects of learning and deterioration. Omega 2007, 35, 397–402. [Google Scholar] [CrossRef]
  11. Yin, Y.; Xu, D.; Sun, K.; Li, H. Some scheduling problems with general position-dependent and time-dependent learning effects. Inf. Sci. 2009, 179, 2416–2425. [Google Scholar] [CrossRef]
  12. Wu, C.-C.; Yin, Y.; Cheng, S.-R. Some single-machine scheduling problems with a truncation learning effect. Comput. Ind. Eng. 2011, 60, 790–795. [Google Scholar] [CrossRef]
  13. Wu, C.-C.; Yin, Y.; Wu, W.-H.; Cheng, S.-R. Some polynomial solvable single-machine scheduling problems with a truncation sum-of-processing-times based learning effect. Eur. J. Ind. Eng. 2012, 6, 441–453. [Google Scholar] [CrossRef]
  14. Wu, C.-C.; Yin, Y.; Cheng, S.-R. Single-machine and two-machine flowshop scheduling problems with truncated position-based learning functions. J. Oper. Res. Soc. 2013, 64, 147–156. [Google Scholar] [CrossRef]
  15. Wang, J.B.; Wang, J.-J. Single machine scheduling with sum-of-logarithm- processing-times based and position based learning effects. Optim. Lett. 2014, 8, 971–982. [Google Scholar] [CrossRef]
  16. Niu, Y.-P.; Wan, L.; Wang, J.-B. A Note on Scheduling Jobs with Extended Sum-of-Processing-Times-Based and Position-Based Learning Effect. Asia Pac. J. Oper. Res. 2015, 32, 1550001. [Google Scholar] [CrossRef]
  17. Zhang, X.; Liu, S.C.; Yin, Y.; Wu, C.-C. Single-machine scheduling problems with a learning effect matrix. Iranian Journal of Science and Technology. Trans. A Sci. 2018, 42, 1327–1335. [Google Scholar]
  18. Biskup, D. A state-of-the-art review on scheduling with learning effect. Eur. J. Oper. Res. 2008, 188, 315–329. [Google Scholar] [CrossRef]
  19. Janiak, A.; Krysiak, T.; Trela, R. Scheduling problems with learning and ageing effects: A survey. Decis. Mak. Manuf. 2011, 5, 19–36. [Google Scholar] [CrossRef]
  20. Azzouz, A.; Ennigrou, M.; Ben Said, L. Scheduling problems under learning effects: Classification and cartography. Int. J. Prod. Res. 2018, 56, 1642–1661. [Google Scholar] [CrossRef]
  21. Wang, J.B.; Xia, Z.Q. Flow-shop scheduling with a learning effect. J. Oper. Res. Soc. 2005, 56, 1325–1330. [Google Scholar] [CrossRef]
  22. Xu, Z.; Sun, L.; Gong, J. Worst-case analysis for flow shop scheduling with a learning effect. Int. J. Prod. Econ. 2008, 113, 748–753. [Google Scholar] [CrossRef]
  23. Wu, C.C.; Wu, W.H.; Hsu, P.H.; Lai, K. A two-machine flowshop scheduling problem with a truction sunm of processing-times-based learning function. Appl. Math. Model. 2012, 36, 5001–5014. [Google Scholar] [CrossRef]
  24. Wang, J.B.; Wang, M.Z. Worst-Case analysis for flow shop scheduling problems with an exponential learning effect. J. Oper. Res. Soc. 2012, 63, 130–137. [Google Scholar] [CrossRef]
  25. Cheng, T.C.E.; Wu, C.C.; Chen, J.C.; Wu, W.H.; Cheng, S.R. Two-machine flowshop scheduling with a truncated learning function to minimize the makespan. Int. J. Prod. Econ. 2013, 141, 79–86. [Google Scholar] [CrossRef]
  26. Wang, X.Y.; Zhou, Z.; Zhang, X.; Ji, P.; Wang, J.B. Several flowshop scheduling problems with truncated position-based learning effect. Comput. Oper. Res. 2013, 40, 2906–2929. [Google Scholar] [CrossRef]
  27. Wang, J.B.; Liu, F.; Wang, J.J. Research on m-machine flow shop scheduling with truncated learning effects. Int. Trans. Oper. Res. 2019, 26, 1135–1151. [Google Scholar] [CrossRef]
  28. Hsu, C.-L.; Lin, W.-C.; Duan, L.; Liao, J.-R.; Wu, C.-C.; Chen, J.-H. A robust two-machine flow-shop scheduling model with scenario-dependent processing times. Discret. Dyn. Nat. Soc. 2020, 2020, 3530701. [Google Scholar] [CrossRef]
  29. Ho, M.H.; Hnaien, F.; Dugardin, F. Electricity cost minimisation for optimal makespan solution in flow shop scheduling under time-of-use tariffs. Int. J. Prod. Res. 2021, 59, 1041–1067. [Google Scholar] [CrossRef]
  30. Lo, T.C.; Lin, B.M.T. Relocation scheduling in a two-machine flow shop with resource recycling operations. Mathematics 2021, 9, 1527. [Google Scholar] [CrossRef]
  31. Chen, X.; Miao, Q.; Lin, B.M.T.; Sterna, M.; Blazewicz, J. Two-machine flow shop scheduling with a common due date to maximize total early work. Eur. J. Oper. Res. 2022, 300, 504–511. [Google Scholar] [CrossRef]
  32. Choi, B.C.; Park, M.J. Two machine fowshop scheduling with convex resource consumption functions. Optim. Lett. 2023, 17, 1241–1259. [Google Scholar] [CrossRef]
  33. Johnson, S.M. Optimal two- and three-stage production schedules with setup times. Nav. Res. Logist. Q. 1954, 1, 61–68. [Google Scholar] [CrossRef]
  34. Holland, J. Adaptation in Natural and Artificial Systems; University of Michigan Press: Ann Arbor, MI, USA, 1975. [Google Scholar]
  35. Essafi, I.; Matib, Y.; Dauzere-Peres, S. A genetic local search algorithm for minimizing total weighted tardiness in the job-shop scheduling problem. Comput. Oper. Res. 2008, 35, 2599–2616. [Google Scholar] [CrossRef]
  36. Nguyen, S.; Mei, Y.; Zhang, M. Genetic programming for production scheduling: A survey with a unified framework. Complex Intell. Syst. 2017, 3, 41–66. [Google Scholar] [CrossRef]
  37. Fan, J.; Zhang, C.; Liu, Q.; Shen, W.; Gao, L. An improved genetic algorithm for flexible job shop scheduling problem considering reconfigurable machine tools with limited auxiliary modules. J. Manuf. Syst. 2022, 62, 650–667. [Google Scholar] [CrossRef]
  38. Tutumlu, B.; Saraç, T. A MIP model and a hybrid genetic algorithm for flexible job-shop scheduling problem with job-splitting. Comput. Oper. Res. 2023, 155, 10622. [Google Scholar] [CrossRef]
  39. Wang, Y.C.; Chen, T. Adapted techniques of explainable artificial intelligence for explaining genetic algorithms on the example of job scheduling. Expert Syst. Appl. 2023, 234 Pt A, 1213769. [Google Scholar] [CrossRef]
  40. Iyer, S.K.; Saxena, B.S. Improved memetic genetic algorithm for the permutation flowshop scheduling problem. Comput. Oper. Res. 2004, 31, 593–606. [Google Scholar] [CrossRef]
  41. Wu, C.-C.; Wu, W.-H.; Chen, J.C.; Yin, Y.; Wu, W.-H. A study of the single-machine two-agent scheduling problem with release times. Appl. Soft Comput. 2013, 13, 998–1006. [Google Scholar] [CrossRef]
  42. Larranaga, P.; Kuijpers, C.M.H.; Murga, R.H.; Inza, I.; Dizdarevic, S. Memetic genetic algorithms for the Travelling Salesman Problem: A Review of Representations and Operators. Artif. Intell. Rev. 1999, 13, 129–170. [Google Scholar] [CrossRef]
  43. Kellegoz, T.; Toklu, B.; Wilson, J. Comparing efficiencies of genetic crossover operators for one machine total weighted tardiness problem. Appl. Math. Comput. 2008, 199, 590–598. [Google Scholar] [CrossRef]
  44. Castelli, M.; Cattaneo, G.; Manzoni, L.; Vanneschi, L. A distance between populations for n-points crossover in memetic genetic algorithms. Swarm Evol. Comput. 2019, 44, 636–645. [Google Scholar] [CrossRef]
  45. Nearchou, A.C. The effect of various operators on the genetic search for large scheduling problems. Int. J. Prod. Econ. 2004, 88, 191–203. [Google Scholar] [CrossRef]
Figure 1. The parameter exploration for GA on n = 10 and 400.
Figure 1. The parameter exploration for GA on n = 10 and 400.
Mathematics 11 04060 g001
Figure 2. Violin plots of AEP for 4 heuristics and GA for small n.
Figure 2. Violin plots of AEP for 4 heuristics and GA for small n.
Mathematics 11 04060 g002
Figure 3. Violin plots of CPU times for 4 heuristics and GA for small n.
Figure 3. Violin plots of CPU times for 4 heuristics and GA for small n.
Mathematics 11 04060 g003
Figure 4. The box plots of RPD for 4 heuristics and GA for big n.
Figure 4. The box plots of RPD for 4 heuristics and GA for big n.
Mathematics 11 04060 g004
Figure 5. The box plots of CPU times for 4 heuristics and GA for big n.
Figure 5. The box plots of CPU times for 4 heuristics and GA for big n.
Mathematics 11 04060 g005
Table 1. Some published recent flow-shop scheduling papers on a two-machine setting.
Table 1. Some published recent flow-shop scheduling papers on a two-machine setting.
ReferencesModelParametersObjective FormAlgorithm
Hsu et al. [28]Two-machine
flow-shop with scenario-dependent processing times
P i j ( s ) ,  i = 1,2;
j = 1 , , n , s = 1 , 2 ,
m i n σ Π m a x s = 1,2 j = 1 n C 2 [ j ] ( s ) ( σ ) A branch-and-bound; a cloud theory-based simulated annealing
Ho et al. [29]Two-machine
flow-shop under time-of-use tariffs
P i j ,  i = 1,2;
j = 1 , , n ,
Min C 2 [ n ] σ or C m a x σ Three heuristics based on Johnson’s rule and three heuristics based on Hadda’s algorithm
Lo and Lin [30]Two-machine flow-shop with resource recycling operations P i j , i = 1,2;
j = 1 , , n ,
Min C 2 [ n ] σ or C m a x σ An integer programming model; ant colony optimization
Chen et al. [31]Two-machine flow-shop with a common due date P i j ,  i = 1,2;
j = 1 , , n ,
Maximize total early workA fully polynomial time approximation scheme
Choi and Park [32]Two-machine flow-shop with convex resource P i j u i j = ( w i j / u i j ) k ,  i=1,2; j = 1 , , n ; w i j > 0, k   > 0.Minimize the sum of the makespan and the total resource consumption costApproximation method
This paperTwo-machine flow-shop with learning dates P i j = { u i j , s i < L D i j v i j , s i L D i j
i = 1,2; j = 1 , , n ,
Min C 2 [ n ] σ or C m a x σ A branch-and-bound; heuristics; population-based GA
Table 2. The performance of four heuristics and GA on the small number of jobs.
Table 2. The performance of four heuristics and GA on the small number of jobs.
n w 1 w 2 BBJ2LD-piJ2U-piJ2V-piJ2UV-piGA
meanmaxmeansdmeansdmeansdmeansdmeansd
80.250.25101469.417.438.657.216.416.555.564.760.010.08
0.50129876497.216.116.635.485.405.365.224.700.030.18
0.755075195944.374.664.274.843.544.163.274.280.000.03
0.500.2520631716.825.337.085.866.715.725.865.020.070.37
0.5089.6310257.565.307.455.025.854.405.334.060.040.19
0.75372719,8754.715.254.784.873.533.663.633.520.030.23
0.750.25146611,4815.304.765.594.824.775.14.394.850.040.30
0.50109080646.695.126.174.565.824.345.074.260.060.26
0.75261012,3314.704.514.934.244.504.044.083.970.050.28
100.250.25483428.006.275.837.135.617.095.925.535.120.210.75
0.50249941,120.009.775.8410.126.376.64.445.824.550.170.41
0.7549,700424,4967.935.378.545.785.654.695.384.470.130.48
0.500.25475291,5328.075.587.604.708.025.416.574.240.280.77
0.50258044,9318.905.099.004.837.654.486.874.410.541.04
0.7544,392486,9877.064.527.564.785.684.574.903.790.220.69
0.750.2576,650533,0204.974.255.374.795.424.574.824.490.180.53
0.5038,918387,6216.564.755.754.625.674.405.594.410.230.64
0.7531,791176,3966.814.926.184.355.583.835.564.110.230.62
120.250.2512911,3666.764.896.985.077.135.506.194.990.210.55
0.50170,3537,713,81110.745.7111.275.598.315.177.435.090.520.99
0.755,014,307776991207.875.487.454.915.964.585.893.950.190.57
0.500.25122,3031,916,4268.625.068.054.738.895.107.734.390.611.15
0.5048,5872,792,14810.095.739.174.608.494.087.974.010.810.96
0.751,363,55516,993,6488.544.617.705.016.494.526.344.170.390.74
0.750.253,589,24732,696,6245.704.885.524.046.455.585.594.730.370.84
0.501,590,39919,702,0246.884.956.905.066.084.445.264.140.530.93
0.751,941,57443,576,6806.974.216.684.746.623.506.314.440.490.99
mean 522494 7.23 7.13 6.235.63 0.25
Table 3. The performance of four heuristics and GA on the big number of jobs.
Table 3. The performance of four heuristics and GA on the big number of jobs.
n w 1 w 2 J2LD_piJ2U-piJ2V-piJ2UV-piGA
meansdmeansdmeansdmeansdmeansd
1000.250.255.913.185.942.716.862.646.162.50.030.21
0.513.263.4411.53.6710.543.348.872.760.000.00
0.7510.964.749.824.359.244.067.563.660.000.00
0.500.2511.043.4410.12.9712.723.1011.723.830.000.00
0.59.053.657.742.688.323.037.712.390.000.00
0.7512.004.149.684.058.253.977.583.370.000.00
0.750.259.714.258.253.5912.094.769.693.860.000.00
0.59.764.977.313.2510.64.229.334.810.000.00
0.757.423.185.973.134.812.835.192.970.000.00
2000.250.252.581.892.291.843.332.002.991.880.410.77
0.510.962.9510.002.858.643.0006.612.660.000.00
0.7510.333.8010.813.869.593.228.162.690.000.00
0.500.259.882.818.572.4811.512.5610.242.650.000.00
0.56.333.244.152.015.372.264.491.950.030.15
0.7511.914.1210.14.488.533.127.82.630.010.05
0.750.2511.395.497.213.0412.134.199.993.970.000.00
0.511.024.887.763.3611.984.729.083.640.000.00
0.759.084.535.873.348.303.056.202.620.010.13
4000.250.251.151.381.031.091.671.61.711.522.481.14
0.506.911.966.181.865.191.872.831.510.010.06
0.7510.093.5310.083.827.732.576.022.560.000.00
0.500.253.921.943.211.676.201.824.671.550.060.32
0.503.632.211.441.213.921.632.351.380.290.59
0.7510.523.919.523.166.982.565.542.280.020.19
0.750.258.363.526.913.0812.493.419.313.080.000.03
0.509.664.366.472.9711.873.858.923.810.010.07
0.756.715.303.492.956.163.344.092.590.370.74
mean 8.65 7.01 8.34 6.84 0.14
Table 4. The summary performance of B&B.
Table 4. The summary performance of B&B.
NodeCPU-Time
nw1meanmean
80.2563850.0242
0.5040230.0117
0.7551670.0124
w2
0.2516840.0060
0.5024780.0061
0.7511,4130.0362
w1
100.2552,2480.1625
0.5051,7250.1832
0.75147,3600.5204
w2
0.2581,4510.2823
0.5043,9970.1530
0.75125,8840.4308
w1
120.255,184,78925.1863
0.501,534,4467.4038
0.757,121,22134.8778
w2
0.253,711,68017.6078
0.51,809,3418.8510
0.758,319,43741.0091
Table 5. The summary AEP performance of proposed algorithms for small n.
Table 5. The summary AEP performance of proposed algorithms for small n.
J2LD-piJ2U-piJ2V-piJ2UV-piGA
nw1meanmeanmeanmeanmean
80.257.006.525.124.680.01
0.506.366.445.364.940.05
0.755.565.565.034.510.05
w2
0.257.187.115.965.270.04
0.507.156.755.695.210.04
0.754.594.663.863.660.03
w1
100.257.998.606.455.580.17
0.508.018.057.126.110.35
0.756.115.775.565.320.21
w2
0.256.446.706.845.640.22
0.508.418.296.646.090.31
0.757.277.435.645.280.19
w1
120.258.468.577.136.500.31
0.509.088.317.967.350.60
0.756.526.376.385.720.50
w2
0.257.036.857.496.500.44
0.59.249.117.636.890.62
0.757.797.286.366.180.36
n
86.316.175.174.710.04
107.377.476.375.670.24
128.027.757.166.520.47
mean7.237.136.235.640.25
Table 6. The summary RPD performance of proposed algorithms for big n.
Table 6. The summary RPD performance of proposed algorithms for big n.
J2LD-piJ2U-piJ2V-piJ2UV-piGA
nw1meanmeanmeanmeanmean
1000.2510.049.098.887.530.01
0.5010.79.179.769.000.00
0.758.967.189.178.070.00
w2
0.258.898.110.569.190.01
0.5010.698.859.828.640.00
0.7510.138.497.436.780.00
w1
2000.257.967.77.195.920.14
0.509.377.618.477.510.01
0.7510.56.9510.88.420.00
w2
0.257.956.028.997.740.14
0.509.447.38.666.730.01
0.7510.448.938.817.390.01
w1
4000.250.256.055.764.863.52
0.500.506.024.725.74.19
0.750.758.245.6210.177.44
w2
0.250.254.483.726.795.23
0.50.56.734.706.994.70
0.750.759.117.706.965.22
n
1009.908.489.278.200.00
2009.287.428.827.280.05
4006.775.376.915.050.36
mean8.657.098.336.840.14
Table 7. The sum of ranks for four heuristics and GA algorithm.
Table 7. The sum of ranks for four heuristics and GA algorithm.
Heuristic/AlgorithmNo. of Obs./Small nRank-Sum (Ri, i = 1, 2,…, 5)No. of Obs./Large nRank-Sum (Ri, i = 1, 2, …, 5)
J2LD-pi2711827112
J2U-pi271172776
J2V-pi278827108
J2UV-pi27552778
GA27272731
The WNMT test indicates that two algorithms are significantly different if |Ri-Rj| > 32, for both small n and large n approximately.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, J.-Y.; Lin, W.-C.; Chang, Y.-W.; Chung, Y.-H.; Chen, J.-H.; Wu, C.-C. A Two-Machine Learning Date Flow-Shop Scheduling Problem with Heuristics and Population-Based GA to Minimize the Makespan. Mathematics 2023, 11, 4060. https://doi.org/10.3390/math11194060

AMA Style

Xu J-Y, Lin W-C, Chang Y-W, Chung Y-H, Chen J-H, Wu C-C. A Two-Machine Learning Date Flow-Shop Scheduling Problem with Heuristics and Population-Based GA to Minimize the Makespan. Mathematics. 2023; 11(19):4060. https://doi.org/10.3390/math11194060

Chicago/Turabian Style

Xu, Jian-You, Win-Chin Lin, Yu-Wei Chang, Yu-Hsiang Chung, Juin-Han Chen, and Chin-Chia Wu. 2023. "A Two-Machine Learning Date Flow-Shop Scheduling Problem with Heuristics and Population-Based GA to Minimize the Makespan" Mathematics 11, no. 19: 4060. https://doi.org/10.3390/math11194060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop