Next Article in Journal
A Real-Time Traffic Sign Recognition Method Using a New Attention-Based Deep Convolutional Neural Network for Smart Vehicles
Next Article in Special Issue
Pharmacophore-Modeling-Based Optimal Placement and Sizing of Large-Scale Energy Storage Stations in a Power System including Wind Farms
Previous Article in Journal
Leg-Joint Angle Estimation from a Single Inertial Sensor Attached to Various Lower-Body Links during Walking Motion
Previous Article in Special Issue
A Survey on Search Strategy of Evolutionary Multi-Objective Optimization Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transfer Learning Based on Clustering Difference for Dynamic Multi-Objective Optimization

School of Computer Science and Technology, Ocean University of China, Qingdao 266100, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(8), 4795; https://doi.org/10.3390/app13084795
Submission received: 28 February 2023 / Revised: 30 March 2023 / Accepted: 4 April 2023 / Published: 11 April 2023
(This article belongs to the Special Issue Advances in Natural Computing: Methods and Application)

Abstract

:
Dynamic multi-objective optimization problems (DMOPs) have become a research hotspot in engineering optimization, because their objective functions, constraints, or parameters may change over time, while quickly and accurately tracking the changing Pareto optimal set (POS) during the optimization process. Therefore, solving dynamic multi-objective optimization problems presents great challenges. In recent years, transfer learning has been proved to be one of the effective means to solve dynamic multi-objective optimization problems. However, this paper proposes a new transfer learning method based on clustering difference to solve DMOPs (TCD-DMOEA). Different from the existing methods, it uses the clustering difference strategy to optimize the population quality and reduce the data difference between the target domain and the source domain. On this basis, transfer learning technology is used to accelerate the construction of initialization population. The advantage of the TCD-DMOEA method is that it reduces the possibility of negative transfer and improves the performance of the algorithm by improving the similarity between the source domain and the target domain. Experimental results show that compared with several advanced dynamic multi-objective optimization algorithms based on different benchmark problems, the proposed TCD-DMOEA method can significantly improve the quality of the solution and the convergence speed.

1. Introduction

Dynamic multi-objective optimization problems (DMOPs) [1,2] are optimization problems in which the objective function and decision variables are related to time (environment). Their optimal solution is a set of Pareto optimal solutions that change dynamically with time (environment). Different from solving the static multi-objective optimization problem, when dealing with this kind of optimization problem, it is necessary to not only optimize several conflicting objectives, but also deal with the changes of objective function and constraints at the same time. DMOPs are widely used to solve many real-world problems, and common application domains include scheduling [3,4], control [5], chemistry [6], industry [7] and energy design [8].
An ideal dynamic multi-objective optimization algorithm should contain three necessary parts, namely change detection, change response and multi-objective optimization algorithm (MOEA) [9,10,11,12]. When the time variable changes, the algorithm needs to detect the change of the objective function in time and respond to the change according to different types of changes, deal with the optimization problem after the change, and use the multi-objective optimization algorithm to iterate the population, then quickly find DPF and DPS at the current moment. In fact, DMOPs can be regarded as a static multi-objective optimization problem (SMOEA) under a set of discrete time variables that can convert complex dynamic characteristics into static processing, which is more convenient and easier to handle, but its disadvantages are also obvious. That is, the processing speed is slow, and timeliness cannot be guaranteed. When encountering an environment with a high frequency of changes, it often fails to achieve the purpose of quickly tracking the Pareto front, resulting in poor performance of the algorithm. Therefore, for ideal DMOEAs, a suitable change response mechanism is the top priority in dealing with DMOPs. The existing environmental response strategies can be divided into the following four categories: diversity maintenance, memory strategy, prediction mechanism and transfer learning [13,14].
Diversity maintenance improves algorithm performance mainly by maintaining or improving population diversity. The specific process is shown in Figure 1a. Xu et al. [15] used the perturbation method to divide the decision variables into time-dependent and time-independent according to the dependence of the decision variables on the time parameters, and respectively adopted the optimal solution of the corresponding decision variables of the co-evolution of two subpopulations. In addition, Zhang et al. [16] maintained population diversity by simulating magnetic particles, and then quickly converged to the Pareto front in the current environment. Generally speaking, the diversity maintenance strategy directly adopts the Pareto optimal solution set of the optimization problem in the previous environment as the initial population in the current environment. Liu et al. [17] used an additional auxiliary strategy to maintain diversity that maintains two archives focusing on convergence and diversity, respectively. In addition, for some problems with biased characteristics, an interval mapping strategy is designed to make their solutions have good diversity. Based on this, Liang et al. [18] divided the decision variables into three parts and adopted the methods of diversity maintenance, prediction and diversity introduction, respectively, to generate high-quality offspring individuals and speed up the convergence of the population. Diversity preservation methods have better performance for DMOPs with weaker changes, but when the optimal solution of the historical environment deviates far from the real Pareto front in the current environment, it will lead to poor problem tracking performance.
Methods based on memory strategies and predictive mechanisms are common approaches to address DMOPs; the process is shown in Figure 1b. Chen et al. [19] proposed an evolutionary algorithm for dealing with time-varying constraints and objective functions. The algorithm employs new mating selection and environment selection operators that allow the population to contain both feasible and infeasible solutions and reuse previous solutions based on information obtained from new environments. Yu et al. [20] used polynomial regression predictors to extract linear or non-linear relationships in historical changes to generate good initial populations for new environments. Zou et al. [21] developed a reinforcement learning method to respond to environmental changes according to the severity of the change, which is considered as three states (mild, moderate and severe). The method adopts three actions of knee-based prediction, center-based prediction and local search, and selects a series of actions according to a given state to reorient the population to a new Pareto front (PF).
Based on the evolution information of the optimization problem in the historical environment, the prediction strategy predicts the fitness terrain or the dominant evolution direction of the current environment and provides useful guidance for the evolutionary optimization process, thereby improving the performance of the algorithm. The process of the forecasting model is shown in Figure 1c. Geng et al. [22] designed a group prediction strategy by converting the individual positions at different moments in the same direction of convergence in the target space into time series, predicting the position at the next moment, improving the diversity and effectiveness of the predicted population, and effectively reducing the convergence time of the algorithm after changing the problem. Wang et al. [23] proposed a prediction strategy based on ensemble learning. The strategy has three forecasting models, including linear and non-linear. Given this, Rong et al. [24] constructed a multi model prediction method for the characteristic change types of translation, rotation and composite problems. Compared with the existing three prediction methods, the proposed prediction mechanism has significant advantages in solving most DMOPs. Zheng et al. [25] used different prediction strategies for different decision variables to generate a new initial population according to the different effects of decision variables on convergence and distribution. Ma et al. [26] recently proposed a feature information prediction algorithm for DMOPs. Among them, the joint distribution adaptation model is used to identify the distribution of solutions after environmental changes and create a population in a new environment on this basis.
Although the prediction model is more widely applicable, it inevitably has prediction errors, which affect the accurate guidance of the optimization process. From a statistical point of view, the solution set used to construct the prediction model and the solution set predicted by the prediction model must obey the independent and identical distribution assumption, thus ignoring the non-independence and identical distribution of the data. In view of this, Jiang et al. [27] introduced the transfer learning strategy as the environment response strategy.
Transfer learning makes full use of the problem information with similar characteristics to guide the prediction or classification of current problems and improve recognition accuracy. Jiang et al. [28] proposed a multi-objective dynamic learning method based on evolutionary learning. Using the transfer principal component analysis method, we learn the Pareto optimal solution set in the adjacent historical environment and generate a set of initial populations through the transfer model in the current environment. Experimental results show that this method can accelerate population convergence and accurately track the optimal solution in the new environment. Jiang et al. [29] proposed a fast dynamic multi-objective evolutionary algorithm based on manifold learning. This method combines the memory mechanism with the learning characteristics of manifold transfer to predict the optimal individual in new instances in the process of evolution. The process is shown in Figure 1d. Liu et al. [30] recently proposed the combination of PPS method and transfer learning to improve population prediction. Jiang et al. [31] proposed a method based on individual transfer learning to solve DMOPs. Unlike the existing method, this method uses a pre-search strategy to filter out some high-quality individuals with good diversity, avoiding negative migration caused by individual agglomeration. Fan et al. [32] also applied transfer learning to solve DMOPs with a large amount of computing, and used alternative auxiliary evolution algorithms, especially MOEA/DEGO, as the baseline to evolve and optimize under limited functional assessment. In addition, transfer learning is used to map the previously archived training data to the current environment to quickly start the proxy model building process, so that the dynamic multi-objective evolutionary algorithm can better adapt to the new environment.
Combining the above transfer learning strategies, the transfer learning strategy has good performance in solving dynamic multi-objective optimization test problems that contain periodic changes and have a large degree of change. However, the application of transfer learning to DMOP solving is still in its infancy. Multiple studies have shown that hybrid change response methods generally perform better than single methods because hybrid methods can handle more diverse dynamic features than single methods. This is demonstrated by the increasing use of mixed strategies in recent work [33].
However, the existing transfer learning methods often need a long training time, which is the main obstacle of some DMOPs. One of the reasons for the slow running speed is that the existing transfer learning methods often realize knowledge reuse by searching the potential space, which will require more parameter settings and consume more computing resources, resulting in a large amount of computing resources wasted on searching low-quality individuals, which greatly increases the possibility of negative transfer [34,35]. If the knowledge possessed by these high-quality individuals can be transferred (from the perspective of convergence and diversity), then more effective and accurate prediction models can be built for the application of DMOPs in various real complex environments.
At the same time, existing dynamic multi-objective optimization algorithms based on forecasting strategies usually use a single forecasting strategy. On the one hand, a single prediction strategy cannot quickly and effectively respond to complex environmental changes; on the other hand, the group diversity generated by a single prediction strategy is poor, and it cannot quickly and effectively track the Pareto front, resulting in the algorithm not being able to quickly converge. Based on the above analysis, in order to reduce the occurrence of negative transfer and improve the running speed, this paper proposes a new environment response mechanism that combines the cluster difference strategy and the transfer learning strategy.
In this paper, the similarity between the source domain and the target domain is improved by adding a clustering difference strategy to predict individuals before transfer learning, thereby reducing the possibility of negative transfer. Then, the transfer learning method TradaBoost [36] is used to build the prediction model. A higher quality initial population is generated through this model, followed by subsequent multi-objective optimization. Therefore, this method is suitable for any population-based multi-objective optimization algorithm and can achieve a large performance improvement. Experiments on different test functions show that the proposed strategy is highly competitive in dealing with problems with different dynamic characteristics and achieves better convergence and distribution.
The contributions of this paper are summarized as follows:
(1)
The Pareto solution set at the next moment is predicted by the clustering difference strategy, so as to narrow the difference between the source domain and the target domain of transfer learning, thereby reducing the possibility of negative transfer. Therefore, the preprocessing process of the target domain is very necessary and can make the subsequent transfer learning more efficient.
(2)
After the target domain is preprocessed, a sample classifier based on the TradaBoost algorithm is used to extract high-quality populations, which can effectively improve the running speed of the algorithm, avoiding more parameter settings and the excessive consumption of computing resources.
The rest of this paper is organized as follows. Section 2 introduces some basic concepts of dynamic multi-objective optimization problems and TradaBoost. Section 3 presents the proposed clustering difference strategy and its combination with transfer learning for solving dynamic multi-objective evolutionary optimization problems. Section 4 presents the experimental setup and results, and discusses the comparison with five other typical dynamic multi-objective optimization algorithms. Section 5 concludes this paper and presents an outlook for future research directions.

2. Background

2.1. Dynamic Multi-Objective Optimization Problems

The mathematical form of dynamic multi-objective optimization of DMOPs is as follows [37]:
M i n i m i z e     F ( x , t ) = f 1 ( x , t ) , f 2 ( x , t ) , , f m ( x , t ) s . t . g i ( x , t ) 0 , i = 1 , 2 , , p h j ( x , t ) = 0 , j = 1 , 2 , , q
where x = x 1 , x 2 , , x n is the decision vector, t is a time or environment variable. f i ( x , t ) : Ω   ( i = 1 , , M ) and Ω = [ L 1 , U 1 × L 2 , U 2 × × L n , U n ] are the lower and upper bounds of the i-decision variable, respectively. g i ( x , t ) 0 , i = 1 , 2 , , p is the i-th inequality constraint, h j ( x , t ) = 0 , j = 1 , 2 , , q is the j-th equality constraint. The purpose of solving the DMOP is to find a set of solutions in different times or environments, so that all objectives are as small as possible.
Definition 1
(Dynamic Decision Vector Domination [38]). At time t, the decision vector x1 Pareto dominates another vector x2, expressed as x 1 x 2 , if and only if
f i ( x 1 , t ) f i ( x 2 , t )         i = 1 , , m f i ( x 1 , t ) f i ( x 2 , t )         i = 1 , , m .
Definition 2
(Dynamic Pareto-Optimal Set (DPS) [39]). If a decision vector x* at time t satisfies
D P S = x * Ω | x Ω , x t x *
For a fixed time window t and a decision vector x * Ω , a decision vector x* is said to be non-dominant if no other decision vector x Ω dominates x*, and the dynamic Pareto-optimal set (DPS) is the set of all non-dominated solutions in the decision space.
Definition 3
(Dynamic Pareto-Optimal Front (DPF) [39]). DPF is the set of the corresponding objective vectors of the DPS, and
D P F = y * | y * = F ( x * , t ) , x * D P S
Algorithm 1 is the main framework of DMOEA. After initializing the current generation of the population, the algorithm employs several strategies to respond to the environment when it changes. The initialized population is updated with the effective strategy, and the time window t is incremented by 1 to represent the next environmental change. In the next step, the i-th multi-objective problem is optimized for generation using a multi-objective evolutionary algorithm. SMOEA uses the updated population as the initial population. Finally, repeat the process if the stopping condition is not met.
Algorithm 1: The main frame of DMOEA.
Input: The number of generations: g; the time window: t;
Output: Optimal solution x* at every time step;
Initialize population P O P 0 ;
While stop criterion is not met do
  if change is detected, then
  Update the population using some strategies: reuse memory, tune parameters, or predict solutions;
  t = t + 1;
  end if
Optimize population with an MOEA for one generation and get optimal solution x*;
end while
g = g + 1;
return x*

2.2. TradaBoost

This paper adopts a method called TradaBoost to meet the requirement of DMOP. TradaBoost is evolved from the Adaboost algorithm, but the Adaboost algorithm, like most traditional machine learning algorithms [40], assumes that the data of the training set and the test set are the same distribution. For migration learning, this assumption is not true. In addition, for the part of the data in the training set that is different from the data in the test set, it will directly lead to a decline in the prediction effect. The TradaBoost algorithm adds weight to each training set sample, and uses the weight to weaken the test set data with different distributions, thereby improving the effect of the model. In each iterative training, if the model misclassifies a source domain sample, then this sample may have a large gap with the target domain sample, so the weight of this sample needs to be reduced. By multiplying the sample by a weight between 0 and 1, through the influence of the weight value, in the next iteration, the influence of this sample on the classification model will be reduced. After a series of iterations, the weights of samples in the source domain that is similar to the target domain or helpful to the classification of the target domain will increase, while the weights of other source domains will decrease.
If there are similarities between multiple source domain datasets and target datasets, in this case, you can try to use multiple source domain datasets to help the learning of the target dataset. More data can be obtained through the above method, so that the relationship between the source data and the target data becomes closer, the transfer process easier, and the classification effect more accurate.
Multi-source TradaBoost assumes that the source training data come from different source domains. In each iteration, the source domain most relevant to the target domain is selected to train a weak classifier, and finally, a strong classifier is obtained. This method can ensure that the transferred knowledge is most relevant to the target task, and through continuous learning, the TradaBoost algorithm can obtain a more accurate classifier for the target domain samples.

3. Proposed TCD-DMOEA

This section details the transfer learning based on clustering difference for the dynamic multi-objective optimization algorithm (TCD-DMOEA). Figure 2 describes the process of TCD-DMOEA. Specifically, first of all, the framework of the algorithm is outlined. Then, the specific process of the clustering type strategy is described. Finally, we analyze the calculation complexity of the strategy.

3.1. Overall Framework

The main framework of TCD-DMOEA is introduced in Algorithm 2. When an environmental change is detected, in the first two changes, evolution is performed using SMOEA. In subsequent changes, the clustering difference strategy is used to process the target domain (see Algorithm 3), and then the transfer learning prediction model (see Algorithm 4) is used to process high-quality individuals as the initial population for subsequent iterations.
Algorithm 2: TCD-DMOEA.
Input: The dynamic optimization problem F t x , a static multi-objective optimization algorithm SMOEA;
Output: The POS of the F t x at the different moments;
Initialization;
P O S t =  SMOEA F t x ;
Generate randomly dominated solutions P t ;
  while the environment has changed, do
  t = t + 1;
  D S t 1 = P O S t 1 P t 1 ;
  D T t = Processing ( F t ( x ) ) ;
  iPop = Case-Transfer ( D S t 1 D T t ) ;
  POSt = SMOEA ( i P o p , F t ( x ) ) ;
  Generate randomly dominated solutions P t ;
  return POSt
end while
Algorithm 3: Processing.
Input: The current population PT; the number of individuals in population, N;
Output: The predicted population PP
Initialize the random population P T and evaluate the initial population PT;
Change detection (PT);
if change is detected, then
  while the maximum number of iterations is not reached, do
    for  i = 1 : N  do
    Use K-means algorithm to cluster the population P into 5 clusters;
    Calculate the centroid C i T of each cluster;
    Calculate C i T + 1 using Formula (8);
    end for
  P T + 1 = C i T + 1 ;
  end while
end if
PP = PT+1;
return PP
Algorithm 4: Case transfer.
Input: The two labeled sets DS and DT, and unlabeled data set D, a based learning algorithm Learner, and the maximum number of iterations N;
Output: The initial population initPop;
Initialize the initial weight vector ω 1 x = 1 D S , x X S 1 D T , x X T ;
  for i = 1     t o     N do
  set P t according to (13);
  Call Learner, providing it the combined training set D with the distribution over D. Then, get back a hypothesis h t : X Y ;
  Calculate ε t according to (9);
  Set β t , Update the weight vector ω i t + 1 according to (10);
  end for
Get h f ( x ) according to (11);
Sample solutions x t e s t at the current environment;
return    i n i t P o p = { x | h f ( x ) = + 1 ,   x x t e s t }

3.2. Processing of Target Domain

The output of the target domain processing stage is the predicted population. The purpose of generating predicted population is to reduce the possibility of negative transfer in subsequent transfer learning. According to the characteristics of transfer learning, negative transfer can be improved by increasing the amount of effective source domain knowledge or reducing the data distribution difference between neighbors. Through the clustering difference strategy, the characteristics of the source domain and the target domain data are first extracted to reduce the difference between the data, and then the knowledge transfer is performed. At the same time, the population quality is improved through the clustering difference strategy to increase the amount of effective source domain knowledge.
In the dynamic environment, the objective function changes with time, but there is a certain relationship between the two objectives before and after the change. Therefore, the optimal solution information before the change can be used to predict the distribution of the next solution. First, the population is divided into five categories using the k-means algorithm, and the centroids of these five categories are calculated separately. Then, the first-order difference is used to predict the next corresponding centroid, and these centroids are formed into a predicted population. Figure 3 shows the process of prediction of clustering differential strategies.
Algorithm 3 gives the pseudo code of target domain processing. The basic principle of K-means algorithm is: assuming a given data sample X, contains n objects X = X 1 , X 2 , X 3 , , X n , each of these objects has m-dimensions attributes. The goal of the K-means algorithm is to cluster n objects into a specified k-class cluster based on similarity between objects. Each object belongs to only one of the class clusters with the smallest distance to the center of the class cluster. For K-means, k cluster centers C 1 , C 2 , C 3 , , C k , 1 < k     n need to be initialized first; the Euclidean distance from each object to the center of each cluster is calculated as shown in the following Formula (5):
d i s ( X i , C j ) = t = 1 m ( X i t C j t ) 2
In the above equation, Xi represents the i-th object 1     i     n , Cj represents the center of the j-th cluster 1     j     k , Xit represents the t-property of the i-th object, 1     t     m , Cjt represents the t-th attribute of the j-th cluster center.
The distance from each object to each cluster center is compared sequentially, and the objects are assigned to the cluster of the nearest cluster center, resulting in k class clusters   S 1 , S 2 , S 3 , , S k .
The K-means algorithm defines the prototype of the class cluster with the center, which is the average of all objects in the class cluster in each dimension, and its calculation process is shown in Formula (6):
C t = X i S l X i S l
where Cl represents the center of the l-th cluster, 1     l     k , Sl represents the number of objects in the l-th class cluster, Xi represents the i-th object in the l-th class cluster, 1     i   S l . The population is divided into five categories according to the K-means principle above, and the centroid C i T i = 1 , 2 , , 5 of each cluster is calculated after clustering.
The first-order differences are then used to derive the centroids C i T i = 1 , 2 , , 5 of each cluster at the next moment. PT is the DPS obtained by the time window T, then C i T can be calculated by the following Formula (7):
C i T = 1 P T x P T x
where C i T + 1 represents the centroid of each cluster in the next time window T+1, as obtained by Formula (8):
C i T + 1 = C i T + C i T C i T 1
where C i T + 1 will constitute a predicted population.

3.3. Transfer Learning

After the predicted population is generated, the source and target domains for transfer learning are specified. Figure 4 describes the instance migration program of TCD-DMOEA. Step 1: Treat the target domain; step 2: The solution in which step 1 is processed and the previous solution (not just the nondominated solutions) are entered into the TradaBoost algorithm. Step 3: Use the TradaBoost algorithm to generate a strong classifier hf. Step 4: Enter hf multiple individuals generated by the current time. Step 5: Individuals recognized as “good” by the classifier hf will form a high-quality initial group.
Here, the TradaBoost algorithm is mainly used to realize the transfer, increase the weight of each training set sample using the weight to weaken the test set data of those different distributions, and then improve the effect of the model. In each iteration of training, if the model misclassifies a sample of the source domain, the sample may have a large gap with the sample of the target domain, so the weight of the sample needs to be reduced. By multiplying the sample by a weight between 0 and 1, by the effect of the weight value, in the next iteration, the effect of the sample on the classification model will be reduced, and after a series of iterations, the weight of the samples in the source domains that are similar to the target domain or helpful to the classification of the target domain will be increased, while the weights of other source domains will be reduced. When the training is complete, the classifier can recognize randomly generated solutions in the current environment and select those individuals identified as “good” by the classifier as the initial population. The specific algorithm pseudo code is described in detail in Algorithm 4.
Algorithm 4 show the main program of the proposed TCD-DMOEA method. During the transfer process, the target domain DT is the predicted population, and the source domain DS is the solution obtained in the past environment. These solutions are then labeled with c ( x ) : x D T D S y , y   { + 1 , 1 } . For each domain, the non-dominant solution is labeled +1, and the dominant solution is labeled −1.
Call Learner, according to the combined training data D and the weight distribution Pt on D and the unlabeled data s, to obtain a classifier h t : X Y in S; statistics on the error rate of ht on Dt by Formula (9):
ε t = i = n + 1 n + m ω i t h t x i c x i i = n + 1 n + m ω i t
If ε t > 0.5 then the algorithm terminates; set up β t = ε t / 1 ε t , β = 1 / 1 + 2 ln n / N , sets the new weight vector by Formula (10):
ω i t + 1 = ω i t β h t x i c x i , i = 1 , , m ω i t β h t x i c x i , i = m + 1 , , m + n
Finally, output the final classifier by Formula (11):
h f ( x ) = 1 , t = N / 2 N ln 1 β t h t x 1 / 2 t = N / 2 N t = N / 2 N ln 1 β t 0 , other
During the pre-build phase of the source domain, the output is the predicted population. In this approach, the main purpose of transfer learning after preprocessing is to reduce the possibility of negative transfer.
Given the sample with an initial weight vector ω 1 = ω 1 1 , , ω n + m 1 , where ω i 1 is obtained from Formula (12):
ω i 1 = 1 m , i = 1 , , m 1 n , i = m + 1 , , m + n
At the same time, set Pt to satisfy Formula (13).
P t = ω t i = 1 n + m ω i t

3.4. Computational Complexity Analysis

This section analyzes the computational complexity of TCD-DMOEA at one iteration. According to Algorithm 2, the main calculation of TCD-DMOEA comes from the following aspects. (1) The complexity of the target domain preprocessing process mainly lies in the use of K-means clustering and first-order difference to predict the non-dominated solution at the next moment. K-means clustering requires O(Inkm) calculation, where m is the number of element fields, n is the amount of data, I represents the number of iterations, and k is the number of clusters. Generally, I, k, and m can be considered as constants, so the computational complexity can be simplified to: O(n). The first difference requires O(n) computation. The computational complexity of the target domain preprocessing stage is O(n). (2) The transfer learning stage uses SVM as the basic classifier, and the SVM classifier costs O(NS2d) to obtain a strong classifier, where S is the overall size, N is the number of iterations, and d is the dimension of the decision variable. To sum up, the computational complexity of TCD-DMOEA in this work is O(NS2d).

4. Experiments

4.1. Test Problems and Performance Indicators

Our experiment was divided into two parts: the first part demonstrated the convergence and distribution uniformity of TCD-DMOEA by comparing it with several popular dynamic multi-objective algorithms. In the second part, through the comparison with Tr-DMOEA [28], it was possible to observe a reduction in the running time, effectively reducing the possibility of negative transfer. The entire experiment was based on MATLAB R2019b, running in a Windows 10 Pro.
The strategy proposed by the experimental setup was mainly used during the initialization stage of the algorithm. A suite of preprocessing and transfer learning techniques allowed us to obtain a high-quality population adapted to the current environment, and evolution to obtain the optimal solution after modification. Theoretically, the target domain generated by the clustering difference strategy is more similar to the source domain, and the initial population improved by the transfer learning method is more adaptable to the changing environment, so as to obtain a solution closer to the actual DPF. To verify the effectiveness of the method, this section used 20 test functions and two related metrics to measure the convergence and uniformity of the algorithm, while using the running time to evaluate the negative transfer possibility of the algorithm.
The 20 test functions used in this section were from CEC 2018: (1) DF function [41], and (2) F function [42]. The DF Benchmark Suite contains 14 questions (DF1-DF14) and the F Benchmark Suite contains six questions (F5–F10). The DF function is a diverse and unbiased benchmark problem, covering various attributes that represent various real scenes, such as time-dependent PF/PS geometry, irregular PF shape, disconnection, knee, etc. F5–F8 in F function have nonlinear correlations among decision variables. The PSs of F5–F7 are 1-D curves, and the PSs of F8 are 2-D surfaces. In F9, the environment changes smoothly in most cases. Sometimes, Pareto sets jump from one region to another. In F10, the geometry of two consecutive PFs is completely different.
In 20 test benchmarks, the time parameter t was used here and defined as t = 1 / n t   τ / τ t , where τ , n t , and τ t represented the maximum number of iterations, the severity of the change, and the frequency of the change, respectively, as described in Table 1. Different kinetic parameters were set for the experiment. We set different variation severities, frequencies of change, and numbers of iterations so that each function could iterate 20 times, and the entire population size was set to 200, meaning that 200 solutions could be generated during evolution. In addition, the k value for K-means was set to 5. This experiment chose RM-MEDA [43] as the SMOEA optimizer of TCD-DMOEA. In this study, the following metrics were used to evaluate the performance of different algorithms.
(1) Inverted generational distance (IGD [44,45,46]): IGD evaluates the convergence and diversity of algorithms by measuring the proximity between the real Pareto frontier and the Pareto frontier obtained by the algorithm, and the definition of the IGD indicator is calculated by Formula (14), where d is calculated by Formula (15).
I G D P F * t , P F t = v P F t * d v , P F t P F t *
d v , P t = min u P t j = 1 m f j v f j u 2
where P F * t is the standard POF of the t-moment, it is the POF obtained by the t-moment algorithm, and d is the Euclidean distance between the individual v on the P F * t and the individual closest to v in P F t . It can be seen that the evaluation method of IGD is for each individual in the standard POF. P F t is used to find the closest point to it in the POF. P F t is obtained by the algorithm, the Euclidean distance between them is calculated, and then all the Euclidean distances are summed and the average taken, so IGD can not only evaluate the proximity between P F * t and PFt, but also evaluate the distribution of individuals in P F t ; the smaller the IGD value, the better the convergence of the POF obtained by the algorithm, and the more uniform the distribution.
The MIGD [42,47] indicator is a variant of IGD and is defined as the average of IGD values over certain time steps in operation. The MIGD value is calculated by Formula (16):
M I G D = 1 T t T I G D ( t )
where IGD(t) represents the IGD value at time t, T is a set of discrete time points in operation, and T is the cardinality of T.
(2) Maximum coverage (MS [31,48]): MS measures the extent to which the Pareto front obtained by the algorithm covers the standard Pareto front. The larger the MS, the better the performance of the algorithm. The MS value is calculated by Formula (17):
M S t = 1 m i = 1 m min P F t i ¯ , P F t i * ¯ max P F t i ¯ , P F t i * ¯ P F t i * ¯ P F t i * ¯ 2
where P F t * is the standard Pareto frontier of t-moment, and P F t is the POF obtained by the t-moment algorithm. Where P F t i ¯ and P F t i ¯ are the maximum and minimum values of the i-th target of the POF obtained by the t-moment algorithm, respectively, P F t i * ¯ and P F t i * ¯ are the maximum and minimum values of the i-th target of the real Pareto frontier at the t-moment.

4.2. Performance Comparison with Other Algorithms

In this section, performance comparison experiments are performed. The above indicators MIGD and MS are used, and some algorithms are compared with the algorithm proposed in this paper. Firstly, the MIGD values of six algorithms are compared, proving the proposed algorithm’s effectiveness. These six algorithms include a dynamic multi-objective optimization algorithm (TCD-DMOEA) that combines clustering difference and transfer learning, a dynamic multi-objective optimization algorithm (Tr-DMOEA) that combines only transfer learning methods [28], a dynamic NSGA-II algorithm (DNSGA-II-A, DNSGA-II-B) [49], PPS [42], and MOEA based on Kalman’s predictions (KF-DMOEA) [50]. A comparison of TCD-DMOEA and Tr-DMOEA was conducted to prove the performance of our proposed strategy. According to the difference in performance index and running time, it could be determined that the clustering difference strategy proposed in this paper can not only maintain good convergence and diversity, but also effectively reduce the possibility of negative transfer.
Figure 5 plots the IGD values obtained by different algorithms after each change. These curves show that under 20 different test functions, the curve obtained by the method was basically at the bottom, and the curve of the method fluctuated less, which meant that the method is not only better performing, but also more stable.
The statistical results of MIGD and MS values that were run 20 times are shown in Table 2 and Table 3, respectively. Table 2 shows the MIGD values for six algorithms in three different configurations. Bold words in the table indicate that the algorithm had the best diversity on this benchmark, and the last column represents the “winner” in this comparison. It can be seen from Table 2 that TCD-DMOEA obtained 44 of the 60 best results, accounting for 73.3%. KF-DMOEA achieved five best results, PPS achieved three best results, and DNSGA-II-A and DNSGA-II-B achieved six and two best results, respectively. Specifically, TCD-DMOEA performed well at most test functions in all dynamic test setups, and DNSGA-II-A achieved better convergence than TCD-DMOEA on DF9 and DF12. The value of PPS was obviously superior to that of other algorithms on DF11.
Table 3 shows the MS values for six comparison algorithms, and it is clear that TCD-DMOEA obtained 38 out of 60 best results, KF-DMOEA obtained only one best result, PPS obtained five best results, and Tr-DMOEA obtained nine best results. DNSGA-II-A and DNSGA-II-B achieved four and three best results, respectively. Specifically, TCD-DMOEA performed well at most test functions in all dynamic test setups, with Tr-DMOEA performing better on F8 and DNSGA-II-B performing better on DF10.
The above experimental results show that TCD-DMOEA could obtain a set of solutions with good convergence and diversity for most test problems. However, it did not perform well enough on some reference functions, such as DF9, DF11, and DF12. TCD-DMOEA also had some shortcomings, which may have been due to the large variation in the POS of these problems, so it was difficult to accurately predict in the target domain processing stage, and the final output of the population quality was poor.
At the same time, it can be found from Table 2 that except for F8, F9 and F10, TCD-DMOEA achieved good performance on the F test function set in most cases. These three test functions were characterized by POS jumps from one region to another, and the geometry of the two consecutive POFs was completely different. This resulted in a low degree of similarity between the source domain and the target domain, which ultimately increased the likelihood of negative transfer.

4.3. Running Speed

The most obvious feature that reduces the possibility of negative transfer is reduced running time, so the running times of different algorithms are compared and shown in Table 4. Table 4 shows that the running time of TCD-DMOEA was the shortest among the six algorithms. This shows that the proposed clustering difference strategy was very effective. TCD-DMOEA ran much faster than Tr-DMOEA. The important difference is that Tr-DMOEA mapped the different distributions that the solution obeys at varying moments to a new potential space through the domain adaptation method, and then found the solution in the potential space, and the construction of the potential space required a huge amount of resources. However, TCD-DMOEA improved the similarity between the source domain and the target domain by preprocessing the population and distinguished the quality of the population through classifiers, which significantly shortened the running time. Figure 6 is a histogram of the running time of several other algorithms except Tr-DMOEA, through which the running time of TCD-DMOEA can be visually compared, and it was found that the running time of TCD-DMOEA was the shortest, while the running times of other algorithms were not much different. The above experimental results show that the proposed TCD-DMOEA method not only improved the running speed of the algorithm, but also greatly improved the quality of the Pareto optimal solution set, resulting in a better state of convergence and distribution for the algorithm.

5. Conclusions

In recent years, transfer learning has been proven to be one of the effective means to solve dynamic multi-objective optimization problems. However, the efficiency of transfer learning (also known as negative transfer) decreases significantly when the target domain is poorly similar to the source domain, or when the transfer learning method is incorrect. Negative transfer forces a search for solutions in the wrong direction, wasting a lot of computing resources. As a result, the running speed becomes slower and the convergence becomes worse.
In this article, a transfer learning method based on a cluster difference dynamic multi-objective optimization algorithm, TCD-DMOEA, was proposed. The TCD-DMOEA applies a clustering difference strategy to increase the similarity between the target domain and the source domain to reduce the likelihood of negative transfer, and the TradaBoost algorithm to classify good-quality populations. Therefore, when the environment changes drastically, the proposed method can improve the quality of the population in the drastically changing environment, thereby improving the convergence and diversity of the dynamic multi-objective algorithm, and at the same time improve the running speed of the algorithm.
The above experimental results fully demonstrate that the algorithm significantly improves the performance of dynamic optimization. Compared with existing transfer learning-based algorithms, the proposed algorithms are tens or even hundreds of times faster at finding POS.
Although the proposed TCD-DMOEA can generate a high-quality initial population, the reliability of the acquired individuals becomes very poor. When the environmental changes are more complex, the accuracy of the cluster-based difference strategy will decrease, and due to the added classifier, the computational complexity of the algorithm will increase. Therefore, in future research work, we will explore the following promising directions. First, it would be beneficial for our static evolution process to try to combine multiple response mechanisms to cope with environmental changes, rather than employing a single strategy. Second, we can try to use classifiers with lower complexity to speed up the optimization process. Additionally, it will be worthwhile to test TCD-DMOEA on a wider range of problems with different types of variation.

Author Contributions

Conceptualization, F.Y.; methodology, F.Y.; software, F.Y.; validation, F.Y.; formal analysis, F.Y.; investigation, F.Y.; resources, F.Y.; data curation, F.Y.; writing—original draft preparation, F.Y.; writing—review and editing, F.Y. and G.-G.W.; visualization, F.Y.; supervision, G.-G.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Key R&D Program of China (Grant No. 2022YFB3305300) and the Fundamental Research Funds for the Central Universities.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are thankful to the anonymous reviewers for their valuable suggestions during the review process.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Parashar, S.; Senthilnath, J.; Yang, X.-S. A novel bat algorithm fuzzy classifier approach for classification problems. Int. J. Artif. Intell. Soft Comput. 2017, 6, 108–128. [Google Scholar] [CrossRef]
  2. Rama, B.; Rosario, G.M. Inventory model with penalty cost and shortage cost using fuzzy numbers. Int. J. Artif. Intell. Soft Comput. 2019, 7, 59–85. [Google Scholar] [CrossRef]
  3. Luo, S.; Zhang, L.; Fan, Y. Dynamic multi-objective scheduling for flexible job shop by deep reinforcement learning. Comput. Ind. Eng. 2021, 159, 107489. [Google Scholar] [CrossRef]
  4. Gao, D.; Wang, G.-G.; Pedrycz, W. Solving fuzzy job-shop scheduling problem using DE algorithm improved by a selection mechanism. IEEE Trans. Fuzzy Syst. 2020, 28, 3265–3275. [Google Scholar] [CrossRef]
  5. Wang, Z.; Ye, K.; Jiang, M.; Yao, J.; Xiong, N.N.; Yen, G.G. Solving hybrid charging strategy electric vehicle based dynamic routing problem via evolutionary multi-objective optimization. Swarm Evol. Comput. 2022, 68, 100975. [Google Scholar] [CrossRef]
  6. Qiao, J.; Zhang, W. Dynamic multi-objective optimization control for wastewater treatment process. Neural Comput. Applications 2018, 29, 1261–1271. [Google Scholar] [CrossRef]
  7. Yang, C.; Ding, J. Constrained dynamic multi-objective evolutionary optimization for operational indices of beneficiation process. J. Intell. Manuf. 2019, 30, 2701–2713. [Google Scholar] [CrossRef]
  8. Barone, G.; Buonomano, A.; Forzano, C.; Palombo, A.; Vicidomini, M. Sustainable energy design of cruise ships through dynamic simulations: Multi-objective optimization for waste heat recovery. Energy Convers. Manag. 2020, 221, 113166. [Google Scholar] [CrossRef]
  9. Wang, G.-G.; Cai, X.; Cui, Z.; Min, G.; Chen, J. High performance computing for cyber physical social systems by using evolutionary multi-objective optimization algorithm. IEEE Trans. Emerg. Top. Comput. 2020, 8, 20–30. [Google Scholar] [CrossRef]
  10. Wang, G.-G.; Gao, D.; Pedrycz, W. Solving multi-objective fuzzy job-shop scheduling problem by a hybrid adaptive differential evolution algorithm. IEEE Trans. Ind. Inform. 2022, 18, 8519–8528. [Google Scholar] [CrossRef]
  11. Wang, G.-G.; Tan, Y. Improving metaheuristic algorithms with information feedback models. IEEE Trans. Cybern. 2019, 49, 542–555. [Google Scholar] [CrossRef] [PubMed]
  12. Yi, J.-H.; Xing, L.-N.; Wang, G.-G.; Dong, J.; Vasilakos, A.V.; Alavi, A.H.; Wang, L. Behavior of crossover operators in NSGA-III for large-scale optimization problems. Inf. Sci. 2020, 509, 470–487. [Google Scholar] [CrossRef]
  13. Cruz, C.; González, J.R.; Pelta, D.A. Optimization in dynamic environments: A survey on problems, methods and measures. Soft Comput. 2011, 15, 1427–1448. [Google Scholar] [CrossRef]
  14. Zheng, R.-Z.; Zhang, Y.; Yang, K. A transfer learning-based particle swarm optimization algorithm for travelling salesman problem. J. Comput. Des. Eng. 2022, 9, 933–948. [Google Scholar] [CrossRef]
  15. Xu, B.; Zhang, Y.; Gong, D.; Guo, Y.; Rong, M. Environment sensitivity-based cooperative co-evolutionary algorithms for dynamic multi-objective optimization. IEEE ACM Trans. Comput. Biol. Bioinform. 2017, 15, 1877–1890. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, K.; Chen, M.; Xu, X.; Yen, G.G. Multi-objective evolution strategy for multimodal multi-objective optimization. Appl. Soft Comput. 2021, 101, 107004. [Google Scholar] [CrossRef]
  17. Liu, Q.; Zou, J.; Yang, S.; Zheng, J. A multiobjective evolutionary algorithm based on decision variable classification for many-objective optimization. Swarm Evol. Comput. 2022, 73, 101108. [Google Scholar] [CrossRef]
  18. Liang, Z.; Wu, T.; Ma, X.; Zhu, Z.; Yang, S. A dynamic multiobjective evolutionary algorithm based on decision variable classification. IEEE Trans. Cybern. 2020, 52, 1602–1615. [Google Scholar] [CrossRef]
  19. Chen, Q.; Ding, J.; Yang, S.; Chai, T. A novel evolutionary algorithm for dynamic constrained multiobjective optimization problems. IEEE Trans. Evol. Comput. 2019, 24, 792–806. [Google Scholar] [CrossRef]
  20. Yu, Q.; Lin, Q.; Zhu, Z.; Wong, K.-C.; Coello, C.A.C. A dynamic multi-objective evolutionary algorithm based on polynomial regression and adaptive clustering. Swarm Evol. Comput. 2022, 71, 101075. [Google Scholar] [CrossRef]
  21. Zou, J.; Li, Q.; Yang, S.; Bai, H.; Zheng, J. A prediction strategy based on center points and knee points for evolutionary dynamic multi-objective optimization. Appl. Soft Comput. 2017, 61, 806–818. [Google Scholar] [CrossRef]
  22. Huan-Tong, G.; Shan-Sheng, Z.; Zhe, C.; Wei-Min, H. Decomposition-based predictive dynamic multi-objective particle swarm optimization algorithm. Control Decis. 2019, 34, 1307–1318. [Google Scholar]
  23. Wang, F.; Li, Y.; Liao, F.; Yan, H. An ensemble learning based prediction strategy for dynamic multi-objective optimization. Appl. Soft Comput. 2020, 96, 106592. [Google Scholar] [CrossRef]
  24. Rong, M.; Gong, D.; Pedrycz, W.; Wang, L. A multimodel prediction method for dynamic multiobjective evolutionary optimization. IEEE Trans. Evol. Comput. 2019, 24, 290–304. [Google Scholar] [CrossRef]
  25. Zheng, J.; Zhou, Y.; Zou, J.; Yang, S.; Ou, J.; Hu, Y. A prediction strategy based on decision variable analysis for dynamic multi-objective optimization. Swarm Evol. Comput. 2021, 60, 100786. [Google Scholar] [CrossRef]
  26. Ma, X.; Yang, J.; Sun, H.; Hu, Z.; Wei, L. Feature information prediction algorithm for dynamic multi-objective optimization problems. Eur. J. Oper. Res. 2021, 295, 965–981. [Google Scholar] [CrossRef]
  27. Jiang, M.; Qiu, L.; Huang, Z.; Yen, G.G. Dynamic multi-objective estimation of distribution algorithm based on domain adaptation and nonparametric estimation. Inf. Sci. 2018, 435, 203–223. [Google Scholar] [CrossRef]
  28. Jiang, M.; Huang, Z.; Qiu, L.; Huang, W.; Yen, G.G. Transfer learning-based dynamic multiobjective optimization algorithms. IEEE Trans. Evol. Comput. 2017, 22, 501–514. [Google Scholar] [CrossRef] [Green Version]
  29. Jiang, M.; Wang, Z.; Qiu, L.; Guo, S.; Gao, X.; Tan, K.C. A fast dynamic evolutionary multiobjective algorithm via manifold transfer learning. IEEE Trans. Cybern. 2020, 51, 3417–3428. [Google Scholar] [CrossRef]
  30. Liu, Z.; Wang, H. Improved population prediction strategy for dynamic multi-objective optimization algorithms using transfer learning. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation (CEC), Kraków, Poland, 28 June–1 July 2021; pp. 103–110. [Google Scholar]
  31. Jiang, M.; Wang, Z.; Guo, S.; Gao, X.; Tan, K.C. Individual-based transfer learning for dynamic multiobjective optimization. IEEE Trans. Cybern. 2020, 51, 4968–4981. [Google Scholar] [CrossRef]
  32. Fan, X.; Li, K.; Tan, K.C. Surrogate assisted evolutionary algorithm based on transfer learning for dynamic expensive multi-objective optimisation problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  33. Hu, Y.; Zheng, J.; Zou, J.; Jiang, S.; Yang, S. Dynamic multi-objective optimization algorithm based decomposition and preference. Inf. Sci. 2021, 571, 175–190. [Google Scholar] [CrossRef]
  34. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, Z.; Dai, Z.; Póczos, B.; Carbonell, J. Characterizing and avoiding negative transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11293–11302. [Google Scholar]
  36. Yao, Y.; Doretto, G. Boosting for transfer learning with multiple sources. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1855–1862. [Google Scholar]
  37. Farina, M.; Deb, K.; Amato, P. Dynamic multiobjective optimization problems: Test cases, approximations, and applications. IEEE Trans. Evol. Comput. 2004, 8, 425–442. [Google Scholar] [CrossRef]
  38. Emmerich, M.T.; Deutz, A.H. A tutorial on multiobjective optimization: Fundamentals and evolutionary methods. Nat. Comput. 2018, 17, 585–609. [Google Scholar] [CrossRef] [Green Version]
  39. Rong, M.; Gong, D.; Zhang, Y.; Jin, Y.; Pedrycz, W. Multidirectional prediction approach for dynamic multiobjective optimization problems. IEEE Trans. Cybern. 2018, 49, 3362–3374. [Google Scholar] [CrossRef]
  40. Cui, Z.; Xue, F.; Cai, X.; Cao, Y.; Wang, G.-G.; Chen, J. Detection of malicious code variants based on deep learning. IEEE Trans. Ind. Inform. 2018, 14, 3187–3196. [Google Scholar] [CrossRef]
  41. Jiang, S.; Yang, S.; Yao, X.; Tan, K.C.; Kaiser, M.; Krasnogor, N. Benchmark Functions for the CEC’2018 Competition on Dynamic Multiobjective Optimization; Newcastle University: Callaghan, Australia, 2018. [Google Scholar]
  42. Zhou, A.; Jin, Y.; Zhang, Q. A population prediction strategy for evolutionary dynamic multiobjective optimization. IEEE Trans. Cybern. 2013, 44, 40–53. [Google Scholar] [CrossRef]
  43. Zhang, Q.; Zhou, A.; Jin, Y. RM-MEDA: A regularity model-based multiobjective estimation of distribution algorithm. IEEE Trans. Evol. Comput. 2008, 12, 41–63. [Google Scholar] [CrossRef] [Green Version]
  44. Van Veldhuizen, D.A.; Lamont, G.B. On measuring multiobjective evolutionary algorithm performance. In Proceedings of the 2000 Congress on Evolutionary Computation. CEC00 (Cat. No. 00TH8512), La Jolla, CA, USA, 16–19 July 2000; pp. 204–211. [Google Scholar]
  45. Ruan, G.; Yu, G.; Zheng, J.; Zou, J.; Yang, S. The effect of diversity maintenance on prediction in dynamic multi-objective optimization. Appl. Soft Comput. 2017, 58, 631–647. [Google Scholar] [CrossRef]
  46. Long, Q.; Li, G.; Jiang, L. A novel solver for multi-objective optimization: Dynamic non-dominated sorting genetic algorithm (DNSGA). Soft Comput. 2022, 26, 725–747. [Google Scholar] [CrossRef]
  47. Liu, M.; Liu, Y. A dynamic evolutionary multi-objective optimization algorithm based on decomposition and adaptive diversity introduction. In Proceedings of the 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Changsha, China, 13–15 August 2016; pp. 235–240. [Google Scholar]
  48. Goh, C.-K.; Tan, K.C. A competitive-cooperative coevolutionary paradigm for dynamic multiobjective optimization. IEEE Trans. Evol. Comput. 2008, 13, 103–127. [Google Scholar]
  49. Deb, K.; Rao, N.U.B.; Karthik, S. Dynamic multi-objective optimization and decision-making using modified NSGA-II: A case study on hydro-thermal power scheduling. In Proceedings of the Evolutionary Multi-Criterion Optimization: 4th International Conference, EMO 2007, Matsushima, Japan, 5–8 March 2007; pp. 803–817. [Google Scholar]
  50. Muruganantham, A.; Tan, K.C.; Vadakkepat, P. Evolutionary dynamic multiobjective optimization via Kalman filter prediction. IEEE Trans. Cybern. 2015, 46, 2862–2873. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Description of environmental response strategies and methods. (a) diversity maintenance. (b) Memory based approach. (c) Prediction based approach (d) Manifold transfer learning method.
Figure 1. Description of environmental response strategies and methods. (a) diversity maintenance. (b) Memory based approach. (c) Prediction based approach (d) Manifold transfer learning method.
Applsci 13 04795 g001
Figure 2. Procedure of TCD-DMOEA.
Figure 2. Procedure of TCD-DMOEA.
Applsci 13 04795 g002
Figure 3. Classification of clustering dynamics.
Figure 3. Classification of clustering dynamics.
Applsci 13 04795 g003
Figure 4. Schematic of procedure case transfer.
Figure 4. Schematic of procedure case transfer.
Applsci 13 04795 g004
Figure 5. IGD values of six algorithms under S2 configuration.
Figure 5. IGD values of six algorithms under S2 configuration.
Applsci 13 04795 g005
Figure 6. Average running time (s) obtained by comparing algorithms under the configuration of n t = 5, τ t = 10.
Figure 6. Average running time (s) obtained by comparing algorithms under the configuration of n t = 5, τ t = 10.
Applsci 13 04795 g006
Table 1. The dynamic parameters.
Table 1. The dynamic parameters.
Settings Change Severity  n t Change Frequency  τ t Maximum Iteration  τ
S1105100
S2510200
S31010200
Table 2. The MIGD values of six comparative algorithms with different dynamic test settings.
Table 2. The MIGD values of six comparative algorithms with different dynamic test settings.
FunctionsSettingsDNSGA-II-ADNSGA-II-BTCD-DMOEAPPSTr-DMOEAKF-DMOEAWinner
S10.125490.184660.018250.799480.167620.18591TCD-DMOEA
DF1S20.141030.149970.026301.007780.226340.20192TCD-DMOEA
S30.069100.070780.017700.348750.174910.15986TCD-DMOEA
S10.079840.110420.005770.568620.10880.12259TCD-DMOEA
DF2S20.098460.098560.005210.550640.18080.14677TCD-DMOEA
S30.043310.052060.005190.331040.13150.10525TCD-DMOEA
S10.360360.314440.061760.460920.3150.38322TCD-DMOEA
DF3S20.313440.300960.227580.564080.41170.36154TCD-DMOEA
S30.342770.353700.058040.281510.34060.37636TCD-DMOEA
S11.335521.296830.927364.090481.56891.20196TCD-DMOEA
DF4S20.850060.851141.453971.780531.83441.98927DNSGA-II-B
S31.299041.286371.034692.136231.58691.98173TCD-DMOEA
S10.098520.185110.023060.365462.31461.35092TCD-DMOEA
DF5S21.675221.681250.023172.002822.62053.26962TCD-DMOEA
S30.057210.072350.022740.256962.47133.30728TCD-DMOEA
S15.507477.964080.4663011.87785.59826.26992TCD-DMOEA
DF6S27.324938.295570.1643612.49867.26928.57652TCD-DMOEA
S32.814713.458560.214335.663126.73796.55881TCD-DMOEA
S14.945948.219140.5193010.08238.84037.45962TCD-DMOEA
DF7S27.437327.842100.3565911.36764.23878.67889TCD-DMOEA
S32.190793.206431.130825.966714.03238.9454TCD-DMOEA
S10.839670.883080.063310.871680.788171.10952TCD-DMOEA
DF8S20.869390.859100.299140.857850.79931.41749TCD-DMOEA
S30.887720.896340.059620.866890.80261.64865TCD-DMOEA
S11.440131.489682.034331.947362.59582.81085DNSGA-II-A
DF9S21.269941.275971.679762.248012.70793.17112DNSGA-II-A
S31.598351.601382.339381.663932.37143.28462DNSGA-II-A
S10.147620.144920.050830.169370.148700.23669TCD-DMOEA
DF10S20.159830.149480.058540.152840.14930.24198TCD-DMOEA
S30.131470.120850.050610.126910.11940.21441TCD-DMOEA
S10.403980.409800.191900.121010.389750.26247PPS
DF11S20.473540.478510.320700.202030.41780.19754PPS
S30.390520.394870.205280.112720.33310.1851PPS
S10.598950.643891.150321.183701.193310.91213DNSGA-II-A
DF12S20.641290.655501.149491.187691.19231.25910DNSGA-II-A
S30.613090.680701.149541.183681.190.989DNSGA-II-A
S10.585720.661710.174890.244563.626203.37829TCD-DMOEA
DF13S22.072382.079510.115301.785612.80323.55246TCD-DMOEA
S30.533170.550640.180220.253702.73121.4413TCD-DMOEA
S10.176730.618800.024710.184371.67272.29643TCD-DMOEA
DF14S22.422692.372780.036351.355171.83332.34678TCD-DMOEA
S30.150270.599570.026260.088401.82571.90653TCD-DMOEA
S11.794442.369320.199015.953152.80264.90052TCD-DMOEA
F5S21.788141.588631.0413814.23583.69195.24772TCD-DMOEA
S30.858221.011560.119642.790082.65926.85162TCD-DMOEA
S11.166221.248860.336712.695811.23494.21603TCD-DMOEA
F6S20.825110.843770.038474.497492.40943.94881TCD-DMOEA
S30.862840.839380.136592.264131.30951.32448TCD-DMOEA
S11.698021.879030.079074.183701.42953.26674TCD-DMOEA
F7S21.534151.570080.0686010.95523.15931.94158TCD-DMOEA
S30.908800.902450.056351.955241.3271.6357TCD-DMOEA
S10.616260.586860.249260.894520.741940.23932KF-DMOEA
F8S20.570030.573020.318960.611231.06150.32661TCD-DMOEA
S30.497230.512840.296690.308420.78750.21715KF-DMOEA
S12.163223.437080.8979916.97651.859470.72251KF-DMOEA
F9S22.792122.791700.2465826.68932.60791.70666TCD-DMOEA
S30.905041.729201.361718.462931.47210.82095KF-DMOEA
S12.804643.231874.1409610.32532.048760.69335KF-DMOEA
F10S21.891472.041390.1719910.36882.78453.83634TCD-DMOEA
S32.589582.524804.105416.394962.73278.66578DNSGA-II-B
Table 3. The MS values of six comparative algorithms with different dynamic test settings.
Table 3. The MS values of six comparative algorithms with different dynamic test settings.
FunctionsSettingsDNSGA-II-ADNSGA-II-BTCD-DMOEAPPSTr-DMOEAKF-DMOEAWinner
S10.872980.833990.995960.652630.92030.7995TCD-DMOEA
DF1S20.874660.871800.996080.645420.841630.7807TCD-DMOEA
S30.920190.925030.993280.854850.889810.8143TCD-DMOEA
S10.919760.902970.997380.528310.94730.8202TCD-DMOEA
DF2S20.914250.912460.997860.719810.916610.8029TCD-DMOEA
S30.949070.944200.988630.746770.939410.833TCD-DMOEA
S10.348430.381870.751880.444200.47010.23TCD-DMOEA
DF3S20.545100.545900.540750.612120.449610.2828PPS
S30.319630.307270.946570.420350.619890.2195TCD-DMOEA
S10.235760.240190.377600.290710.43830.2705Tr-DMOEA
DF4S20.337260.339180.282170.377720.244220.2985PPS
S30.231910.234890.564080.287550.317610.308TCD-DMOEA
S10.995500.996260.999910.9923310.9426Tr-DMOEA
DF5S20.997690.996850.999880.999590.998650.956TCD-DMOEA
S30.997900.998360.999930.999980.999000.9303PPS
S10.890980.949270.999090.93139510.7099Tr-DMOEA
DF6S20.993250.997240.999990.8984780.6328450.8682TCD-DMOEA
S30.965540.980840.999620.9662980.615650.75TCD-DMOEA
S10.90.937850.956600.92077510.7155Tr-DMOEA
DF7S2110.827910.90.6915490.8441DNSGA-II-A
S30.924980.9474310.865270.666240.7515TCD-DMOEA
S10.365730.350960.863520.371470.45010.3004TCD-DMOEA
DF8S20.352440.343340.638000.375480.710350.6078Tr-DMOEA
S30.326560.330390.924330.325420.657140.4329TCD-DMOEA
S10.852040.847530.351280.7639240.80680.6775DNSGA-II-A
DF9S20.914610.929850.696640.929340.741040.7588PPS
S30.764680.787700.831420.809240.646140.7012TCD-DMOEA
S10.9881010.999990.994260.99980.9502DNSGA-II-B
DF10S20.9999910.999990.995590.999380.9927DNSGA-II-B
S30.9915110.999990.998420.999620.9125DNSGA-II-B
S10.718510.710670.946980.946670.97090.9008TCD-DMOEA
DF11S20.694610.693530.860000.9164070.97620.9891KF-DMOEA
S30.724550.721700.94380.961960.998930.9826Tr-DMOEA
S10.496250.476940.001050.064890.00450.3006DNSGA-II-A
DF12S20.532520.528520.002120.001850.000890.106DNSGA-II-A
S30.490160.448080.600660.001740.005680.1009TCD-DMOEA
S10.993240.992420.997940.9956360.9950.9087TCD-DMOEA
DF13S20.994850.990800.997060.998170.997210.9479PPS
S30.994060.994340.997810.997080.996110.9361TCD-DMOEA
S10.926240.788560.949980.925840.9270.7855TCD-DMOEA
DF14S20.768540.752490.949980.821290.903910.826TCD-DMOEA
S30.926490.812700.949960.934500.916220.772TCD-DMOEA
S10.386980.473430.945830.344780.67680.6133TCD-DMOEA
F5S20.455550.480110.867380.456130.670230.7623TCD-DMOEA
S30.579740.587420.943560.584050.599820.6696TCD-DMOEA
S10.529460.502490.960140.464200.69170.5548TCD-DMOEA
F6S20.613630.586330.965020.567260.675530.6745TCD-DMOEA
S30.569280.596850.978480.664600.596920.4971TCD-DMOEA
S10.490830.501850.978990.5589750.64820.5175TCD-DMOEA
F7S20.531200.576480.971980.466780.648960.7871TCD-DMOEA
S30.544820.606650.980710.735430.593420.5414TCD-DMOEA
S10.998650.999410.999980.9992610.9871Tr-DMOEA
F8S20.999770.999910.999980.9997710.998Tr-DMOEA
S30.999080.999930.999970.9999710.9835Tr-DMOEA
S10.442870.338250.899250.297840.68310.5528TCD-DMOEA
F9S20.436990.3339720410.960780.260470.588840.6172TCD-DMOEA
S30.521200.420070.904540.405610.649520.5602TCD-DMOEA
S10.654580.682980.985270.481780.66990.5614TCD-DMOEA
F10S20.523820.464900.957320.506960.694610.7753TCD-DMOEA
S30.706960.739790.998730.509120.770930.6982TCD-DMOEA
Table 4. The average running time (seconds) of the six algorithms under the S2 configuration.
Table 4. The average running time (seconds) of the six algorithms under the S2 configuration.
FunctionsDNSGA-II-ADNSGA-II-BTCD-DMOEAPPSTr-DMOEAKF-DMOEA
DF123.025827.028910.002830.1587461.559130.4620
DF229.856128.41969.897429.7459363.269829.4512
DF339.27430.478510.052518.4148459.265827.1253
DF448.147830.07539.759817.4654165.580223.3695
DF529.479631.75928.889530.6544399.003632.4890
DF631.154829.29718.953624.8569524.201621.0256
DF743.457428.14909.258725.2584378.015423.1016
DF830.476529.02538.125828.5298425.054527.4694
DF930.465436.46958.589120.1489484.015425.9726
DF1024.898924.64978.185689.2103866.596297.6546
DF1121.489025.02459.978281.0365890.416891.6469
DF1226.598221.469410.015949.2016970.146047.4102
DF1319.002321.75927.998785.41601093.000198.7912
DF1418.889819.165412.023885.0489908.459096.4694
F5132.000132.079513.017348.2203513.416959.8591
F663.899431.591113.592450.1605545.241660.4991
F754.656557.225912.365459.2056571.156061.1697
F832.787956.741113.248576.85691269.203686.7952
F958.989139.289912.249852.5892970.259663.7961
F1057.459029.059511.147549.0412597.526952.5610
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, F.; Wang, G.-G. Transfer Learning Based on Clustering Difference for Dynamic Multi-Objective Optimization. Appl. Sci. 2023, 13, 4795. https://doi.org/10.3390/app13084795

AMA Style

Yao F, Wang G-G. Transfer Learning Based on Clustering Difference for Dynamic Multi-Objective Optimization. Applied Sciences. 2023; 13(8):4795. https://doi.org/10.3390/app13084795

Chicago/Turabian Style

Yao, Fangpei, and Gai-Ge Wang. 2023. "Transfer Learning Based on Clustering Difference for Dynamic Multi-Objective Optimization" Applied Sciences 13, no. 8: 4795. https://doi.org/10.3390/app13084795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop