Next Article in Journal
An Adaptive-to-Model Test for Parametric Functional Single-Index Model
Next Article in Special Issue
NSGA-II/SDR-OLS: A Novel Large-Scale Many-Objective Optimization Method Using Opposition-Based Learning and Local Search
Previous Article in Journal
New Solitary Wave Patterns of the Fokas System in Fiber Optics
Previous Article in Special Issue
Active Debris Removal Mission Planning Method Based on Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Learning Moth Search Algorithm for Solving Multidimensional Knapsack Problems

1
School of Information Engineering, Hebei GEO University, Shijiazhuang 050031, China
2
Intelligent Sensor Network Engineering Research Center of Hebei Province, Shijiazhuang 050031, China
3
Xinjiang Institute of Engineering, Information Engineering College, Ürümqi 830023, China
4
Shanwei Institute of Technology, Shanwei 516600, China
5
School of Computer Science and Engineering, Huizhou University, Huizhou 516007, China
*
Authors to whom correspondence should be addressed.
Mathematics 2023, 11(8), 1811; https://doi.org/10.3390/math11081811
Submission received: 8 March 2023 / Revised: 31 March 2023 / Accepted: 4 April 2023 / Published: 11 April 2023
(This article belongs to the Special Issue Evolutionary Computation 2022)

Abstract

:
The moth search algorithm (MS) is a relatively new metaheuristic optimization algorithm which mimics the phototaxis and Lévy flights of moths. Being an NP-hard problem, the 0–1 multidimensional knapsack problem (MKP) is a classical multi-constraint complicated combinatorial optimization problem with numerous applications. In this paper, we present a hybrid learning MS (HLMS) by incorporating two learning mechanisms, global-best harmony search (GHS) learning and Baldwinian learning for solving MKP. (1) GHS learning guides moth individuals to search for more valuable space and the potential dimensional learning uses the difference between two random dimensions to generate a large jump. (2) Baldwinian learning guides moth individuals to change the search space by making full use of the beneficial information of other individuals. Hence, GHS learning mainly provides global exploration and Baldwinian learning works for local exploitation. We demonstrate the competitiveness and effectiveness of the proposed HLMS by conducting extensive experiments on 87 benchmark instances. The experimental results show that the proposed HLMS has better or at least competitive performance against the original MS and some other state-of-the-art metaheuristic algorithms. In addition, the parameter sensitivity of Baldwinian learning is analyzed and two important components of HLMS are investigated to understand their impacts on the performance of the proposed algorithm.

1. Introduction

The multidimensional knapsack problem (MKP) [1,2] is a generalization of the 0–1 knapsack problem (0–1 KP) [3]. As a classical combinatorial optimization problem, numerous real-world applications can be modeled as MKP, such as capital budgeting [4,5], cutting stock [6], and loading problem [7,8].
Then the MKP is: given a set of n items and a set of m knapsacks (m < n), with cj = profit of item j, bi = capacity constraint of knapsack i,aij = resource consumption of item j in the ith knapsack. The goal of MKP is to select a subset of items so that the total profit of the selected items is a maximum while the total weights in each dimension i (i = 1, 2, …, m) do not exceed the corresponding capacity bi. Formally,
m a x z = j = 1 n c j x j
s . t . j = 1 n a i j x j b i , i { 1 , 2 , , m }
x j { 0 , 1 } , j { 1 , 2 , , n }
where xj (j = 1, 2, …, n) is a 0–1 decision variable, such that xj = 1 if item j is assigned to a knapsack, xj = 0 otherwise. When m = 1, MKP reduces to the 0–1 KP.
Known as the NP-hard combinatorial optimization problem of MKP, the conventional exact algorithms usually are unable to obtain a satisfactory solution in a reasonable time, especially for large-scale instances. It is because that combinatorial optimization problems often feature a combination of the explosion and then the search space grows exponentially with the expansion of the scale. In this case, metaheuristic algorithms are effective methods for MKP, which can obtain near-optimal solutions in a reasonable acceptable time. Representative metaheuristic algorithms for solving MKP include the two-phase tabu evolutionary algorithm (TPTEA) [9], binary particle swarm optimization (PSO) [10], quantum particle swarm optimization (QPSO) [11], diversity-preserving quantum particle swarm optimization (DQPSO*) [12], hybrid estimation of distribution algorithm (EDA) [13], memetic algorithm (MA) [14], binary grey wolf optimizer (GWO) [15], hybrid harmony search algorithm (HS) [16], binary multi-verse optimizer (MMVO) [17], cuckoo search (CS) [18], pigeon-inspired optimization algorithm (PIO) [19], binary moth search algorithm (MS) [20], sine cosine algorithm (SCA) [21], and binary slime mould algorithm (BSMA) [22]. For more information on evolutionary algorithms (EAs) [23] and exact methods for solving MKP problem, please refer to [24].
Recently, some novel metaheuristic algorithms have been proposed and used to solve various optimization problems, including continuous optimization problems and discrete optimization problems, such as differential evolution algorithm (DE) [25,26], cuckoo search algorithm (CS) [27], bat algorithm (BA) [28], krill herd (KH) [29,30], elephant herding optimization (EHO) [31], monarch butterfly optimization algorithm (MBO) [32,33], brain storm optimization algorithm (BSO) [34], fruit fly optimizer (FFO) [35], whale optimization algorithm (WOA) [36], wind driven optimization (WDO) [37], salp swarm algorithm (SSA) [38], Harris hawks optimization (HHO) [39], artificial Jellyfish Search (JS) optimizer [40], artificial rabbits optimization (ARO) [41], mountain gazelle optimizer (MGO) [42], moth-flame optimizer (MFO) [43], and many more.
The moth search (MS) [44,45] algorithm was recently developed by Wang and inspired by the phototaxis and Lévy flights of moths. Owing to its simplicity and high search efficiency, MS has been successfully applied to solve various optimization problems, such as the cloud task scheduling problem [46], drone placement problem [47], constrained optimization problems [48], discounted {0–1} knapsack problem [49], and set-union knapsack problem [50]. Previous studies have shown that MS is effective for solving various combinatorial optimization problems, especially for various variants of KP problems [49,51]. To the best of our knowledge, there are few literatures on applying MS to solve MKP. Hence, we concentrated on the MS algorithm for solving MKP.
In recent years, the learning-based metaheuristic algorithms have been extensively reported in literature because of their effectiveness and efficiency in solving various optimization problems. The core idea is that the metaheuristic algorithm combines specific learning operators or learning mechanisms to enhance itself some learning ability, and then owning better optimization behavior. There are many learning methods, for example, deep learning [52,53], reinforcement learning [54,55,56], transfer learning [57], Q-learning [58], information feedback [59], orthogonal learning [34], comprehensive learning [60], Baldwinian learning, and so on.
Inspired by Darwinian evolution, evolutionary computation mainly includes the random variation of individuals and fitness selection mechanism [61]. It seems time-consuming to search randomly for good genotypes without using phenotypes. One of the possible ways to surmount this deficiency is to integrate the learning mechanism into the evolutionary search. The learning mechanism can provide a more effective search path [62]. Hence, learning-based mechanisms have been extensive researched and applied in enhancing the search performances of evolution algorithms [14,31]. The Baldwin effect, sometimes called Baldwinian learning, was widely incorporated into a variety of evolutionary computing models to enhance their performance [61,63,64]. The above works motivate us to propose a Baldwinian learning MS to solve MKP.
Harmony search (HS) [65] is metaheuristic and imitates the musical improvization process to search for a perfect state of harmony. Since proposed, HS has always attracted much attention from researchers and has been successfully applied to deal with various optimization problems. An effective variant of HS, called global-best HS (GHS), was proposed by Omran [66]. Compared with HS and other variants, GHS alleviates the problem of tuning the parameter bw and can work efficiently on both continuous and discrete problems. The GHS algorithm, due to its easy implementation and quick convergence, has been applied to many fields. Xiang et al. [67] proposed a discrete GHS (DGHS) for solving 0–1 KP. Keshtegar et al. [68] proposed a Gaussian GHS (GGHS) algorithm for solving numerical optimization problems. EI-Abd et al. [69] proposed an improved GHS (IGHS) algorithm to solve continuous optimization. In essence, musical performances seek to find pleasing harmony, just like the process of finding the global optimal solution through continuous learning. Based on this, GHS, as another learning scheme, is integrated into the MS for a global search.
Aiming at resolving the above issues, we propose a hybrid learning MS (HLMS), by introducing the Baldwinian learning and GHS learning mechanism. In HLMS, a novel Baldwin learning strategy based on Cauchy distribution is proposed instead of Gaussian distribution in [70]. During the evolutionary process, each individual in the whole population performs GHS learning and Baldwinian learning successively with a certain probability to reduce the time-consumption. The beneficial combination and complementarity of these two mechanisms lead HLMS to evolve towards the global optimum. Intuitively, HLMS has better search performance than the original MS because of the good balance between the exploration capacity of GHS learning and the exploitation ability of the Baldwinian learning.
The novelty and main contributions of this work include.
  • A novel Baldwinian learning strategy based on Cauchy distribution is proposed, the analysis and experiment of which show this strategy is more effective than the Baldwin learning strategy based on Gaussian distribution.
  • Combined with Baldwin learning, GHS is an effective global search operator to enhance the exploration ability of HLMS. Meanwhile, the pitch adjustment process based on dimensional learning can generate a large jump in the search process.
  • To reduce computation costs, the proposed HLMS triggers Baldwin learning and GHS learning with a certain probability in each iteration.
  • Exploration and exploitation are two common and fundamental features of any optimization method. In the evolution of HLMS, Lévy flight and GHS learning are mainly responsible for exploration, whilst flight straightly and Baldwinian mainly implement exploitation.
The rest of this paper is organized as follows. Section 2 reviews the MS algorithm, the Baldwinian learning scheme, and the GHS algorithm. In Section 3, the proposed HLMS for the MKP is introduced in detail. Extensive experiments and comparisons are conducted in Section 4. Finally, conclusions and suggestions are provided in Section 5.

2. Preliminaries

In this section, we summarize the core idea of MS, Baldwinian learning, and Global-best HS algorithm, which form the basis of the proposed HLMS framework.

2.1. Moth Search Algorithm

The moth search algorithm (MS) [44] is a new metaheuristic algorithm developed by Wang and inspired by the phototaxis and Lévy flights of moths. Based on these two behaviors, the flight straightly operator and Lévy flight operator of MS are derived, which can achieve good balance between the exploration capability and exploitation ability. Meanwhile, the whole population is subdivided into two subpopulations (named subpopulation1 and subpopulation2) based on the fitness of moth individuals. To this end, the position of offspring in subpopulation1 and subpopulation2 is updated by Lévy flight and flight straightly, respectively.
In the Lévy flight stage, the core mathematical formulation is:
X i t + 1 = X i t + α L ( s )
α = S m a x / t 2
L ( s ) = ( β 1 ) Γ ( β 1 ) sin ( π ( β 1 ) 2 ) π s β
where Xit and Xit+1 denote the position vector of the ith moth at generation t and t + 1, respectively.  α is the scale factor and Smax is the max walk step. L(s) represents the step drawn from Lévy distribition with  β = 1.5 and  Γ ( x ) is the gamma function.
In the flight straightly stage, the ith individual in subpopulation2 is considered to fly in a straight line towards the light source. The mathematical model of the flight straightly operator is formulated as follows:
X i t + 1 = { λ × ( X i t + φ × ( X b e s t t X i t ) ) i f r a n d > 0.5 λ × ( X i t + 1 φ × ( X b e s t t X i t ) ) e l s e
where  λ is the scale factor, which is used to control the convergence speed of the algorithm and improve population diversity.  λ is set to a random number drawn by the standard uniform distribution. The acceleration factor  φ is set to golden ratio (0.618).  X b e s t t is the best moth individual at generation t. rand returns a random number that is uniformly distributed in (0, 1).
The pseudo code of MS is shown in Algorithm 1.
Algorithm 1. Moth search algorithm
Begin
Step 1: Initialization.
 Set the maximum iteration number MaxGen and iteration counter G = 1; Initialize the parameters max walk step Smax, the index  β , and acceleration factor  φ .
 According to uniform distribution, the population with NP individuals is randomly initialized.
Step 2: Fitness calculation.
 Compute the initial objective function values of each individual according to their position. Memory the best individual (denotes as Xbest).
Step 3: While (G < MaxGen) do
 Divide the whole population into two subpopulations with equal size: subpopulation1 and subpopulation2, based on their fitness.
 Update subpopulation1 by using Lévy flight operator (Equation (4)).
 Update subpopulation2 by using flight straightly operator (Equation (7)).
 Evaluate the objective function values of each individual and update Xbest.
G = G + 1.
 Sort the population by fitness.
Step 4: End while
Step 5: Output: the best results.
End.

2.2. Baldwin Effect and Baldwinian Learning

The interactive way of learning and evolution was first proposed by Baldwin, known as the Baldwin effect. Hence, different Baldwinian learning models have been proposed based on the Baldwin effect. Hinton and Nowlan [62] found that it was difficult to find the optimal solution of more complex problems only by an evolutionary algorithm. However, when combined with Baldwinian learning, the performance of the hybrid algorithm can be effectively improved.
Generally speaking, Baldwinian learning is a type of local search strategy in an evolutionary algorithm. The Baldwinian learning mechanism was first combined with the clonal selection algorithm (CSA) by Gong et al. [61] to improve the performance of BCSA. Based on this, Peng et al. [70] proposed four Baldwinian learning strategies inspired by the trial vector generating strategy of differential evolution (DE).
In evolutionary computation, Cauchy mutation and Gaussian mutation are two popular and effective mutation techniques [71]. The characteristic of Gaussian mutation is to speed up the local convergence, and Cauchy mutation is better at escaping from local optimum. However, compared with Gaussian mutation, Cauchy mutation is insensitive to mutation step size and can achieve the acceptable performance.
According to the above analysis, the newly designed Baldwinian learning operator based on Cauchy distribution is proposed in HLMS. The mathematical expression is as follows:
Y i = X r 1 + c ( X r 2 X r 3 )
where  Y i is the donor vector for each moth individual  X i from the current population after applying Baldwinian learning.  X r 1 X r 2 , and  X r 3 are sampled randomly from the current population and r1, r2, and r3 are mutually exclusive integers randomly chosen from the range [1, NP], which are also different from the selected individual index i. The parameter c is the strength of Baldwinian learning and is a random number based on Cauchy distribution.

2.3. Global-Best HS Algorithm

GHS is a novel variant of HS and inspired by the concept of swarm intelligence of PSO [66]. The difference from the original HS is that the new harmony can mimic the best harmony in the harmony memory HM. Meanwhile, the parameter bw in HS is replaced and a social dimension is added to the GHS. In addition, the GHS dynamically updates the pitch adjusting rate PAR according to the following equation [72]:
P A R ( t ) = P A R m i n + P A R m a x P A R m i n N I × t
where PAR(t) is the pitch adjusting rate for generation t,  P A R m i n is the minimum adjusting rate, and  P A R m a x is the maximum adjusting rate. NI is the number of improvisations and t is the generation number.
The procedure of GHS is given in Algorithm 2.
Algorithm 2. The Global-best HS algorithm (GHS)
Begin
For each  i [ 1 , n ]  do     /*n is the dimension of the problem*/
If U(0, 1) ≤ HMCR then   /* memory consideration*/
x i h = x i j , where  j U ( 1 , , H M S )
If U(0,1) ≤ PAR (t) then /*pitch adjustment*/
x i h = x k b e s t , where best is the best harmony in the HM and  k U ( 1 , n )
Else         /*random selection*/
x i h = L B i + r a n d * ( U B i L B i )
End.
Furthermore, it should be emphasized that, the concept of dimensional learning [73] is embodied in Algorithm 2, as shown below:
X i h = X k b e s t
In Equation (7) of MS, the same dimension j is selected in  X b e s t t X i t for conducting the new solution. Under this dimension, if the component value of the ith individual is similar to the best individual, the difference  X b e s t t X i t will be very small, especially in the later stage of evolution. This means that such a step size is not conductive to  X i jumping to a far position. If the best moth individual is local optimum, the solution hardly escapes from the local extremum. In Equation (10), the dimension index i of  X h is not equal to the dimension index k of  X b e s t . Generally, the difference between two different dimensions is large. Different dimensions can carry different information.
Based on the above analysis, dimension learning is embedding into Algorithm 2 and it should be an effective global search operator.

3. The Proposed HLMS for the MKP

The proposed HLMS algorithm for MKP is inspired from the studies [66,70] and distinguishes itself with two new features. First, GHS as a powerful global search operator is introduced to enhance the exploration ability of the algorithm. Second, a new Baldwinian learning strategy by replacing Gaussian distribution with Cauchy distribution is introduced on HLMS. The proposed algorithm framework and the main components of the MKP problem are described in the following subsection.

3.1. Population Initialization

In this stage, NP moth individuals are randomly generated in the search space. The swarm X = {X(1), X(2), …, X(NP)} is maintained and evolves, where each moth individual X(i) is a n-dimensional real-valued vector  X ( i ) = ( x 1 i , x 2 i , , x n i ) with  x j i { a , a } j { 1 , 2 , , n } and n is the number of objects or items. Here, a takes the value 3 or 5 in this paper. Then, each moth individual X(i) is transformed into an n-dimensional binary vector by a mapping method, which is called discrete moth  Y ( i ) = ( y 1 i , y 2 i , , y n i ) with  y j i { 0 , 1 } j { 1 , 2 , , n } .

3.2. Solution Representation

In HLMS, an n-bit binary string consisting of 0 and 1 is used to represent a candidate solution. If the item is selected, the bit is 1, otherwise it is 0. It should be noted that the MKP is a constrained optimization problem, so the solution generated in the evolution process may be infeasible.
In this paper, a simple and effective transfer function [50] is adopted and the function expression is as follows:
T ( x ) = x
The transfer method from a real-valued variable  x i to a binary variable  y i is calculated by:
y i = { 1 , i f T ( x i ) > 0 0 , else

3.3. Quick Repair Operator

Learning from previous research work [10,74], the HLMS algorithm also adopts a popular quick repair operator based on pseudo-utility which was proposed by Luo et al. [15]. In order to effectively apply the repair operator, the given MKP instance needs to be preprocessed. Specifically, all the items are renumbered in an ascending order based on their scaled pseudo-utility ratios  σ j [75] defined as follows:
σ j = c j i = 1 m a i j b i , j { 1 , 2 , , n }
More exactly, the index values of all sorted items are stored in array J [1…n], such that  σ J [ 1 ] σ J [ 2 ] σ J [ n ] . The vectors  ( c 1 , c 2 , , c n ) and  a i , j (i = 1, 2, …, m, j = 1, 2, …, n) should be adjusted as well. The main framework of the quick repair procedure can be summarized in Algorithm 3.
Algorithm 3. The quick repair operator base on the scaled pseudo-utility
Begin
Step 1: Input.  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
Step 2: Calculation. Calculate the current total consumption of each resource.
r i = j = 1 n a i j * x j , i = 1 , 2 , , m .
Step 3: Repair process.
For j = n to 1 do
If  r i b i , i = 1 , 2 , , m then
 Break.
else if  x J [ j ] = 1 then
 Set  x J [ j ] = 0, and  r i r i a i J [ j ] , i = 1 , 2 , , m .
End
End
Step 4: Optimization process.
For j = 1 to n do
If  x J [ j ] = 0 and  r i + a i J [ j ] b i , i = 1 , 2 , , m  then
 Set  x J [ j ] = 1 and  r i r i + a i J [ j ] , i = 1 , 2 , , m .
End
End
Step 5: Output:  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
End.
Obviously, the quick repair operator of Algorithm 3 mainly consists of two phases. In the first phase, called the repair process, according to the ascending order of the scaled pseudo-utility ratios, the items are removed from the knapsack one by one until the solution is feasible. In the second phase, called the optimization process, for all the feasible solutions, greedily packed the items to be loaded into the knapsack based on the descending order of the scaled pseudo-utility ratios one by one. In this process, the feasibility of the solution needs to be maintained all the time. In brief, the first phase makes all the infeasible solutions become feasible, and the second phase enables the quality of feasible solutions better.

3.4. Procedure of HLMS for MKP

Based on the analysis above, the proposed HLMS algorithm for MKP is outlined in Algorithm 4. The algorithm framework includes the following main steps. (1) After initialization, the repaired population is divided into two subpopulations based on the fitness. (2) Subpopulation1 and subpopulation2 apply the Lévy flight operator and flight straightly operator, respectively. (3) GHS learning and Baldwinian learning are implemented in sequence to the whole population with a probability of 0.5. (4) The mapping from the real-valued vector to the binary vector is realized with a transfer function and then the repair of infeasible solutions and the optimization of feasible solutions are performed. (5) Evaluating the solution is based on the objective function and then the whole population is divided into two subpopulations. Steps (2)(5) are repeated until the termination condition is reached.
Algorithm 4. Procedure of HLMS for MKP
Begin
  •  Step 1: Initialization.
  • Set the maximum iteration number MaxGen and iteration counter G = 1; Initialize the parameters max walk step Smax, the index  β φ , strength of Baldwinian learning c.
  • According to uniform distribution, the population with NP individuals is randomly initialized.
  • The transform function is used to discretize the real number vector to obtain the initial solution  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
  • Repair the initial solution by Algorithm 3.
  • Step 2: Fitness evaluation.
  • Evaluate the initial solution using the objective function of MKP.
  • Step 3: Whileg < MaxGen do
 3.1 Divide the whole population into two subpopulations with equal size: subpopulation1 and subpopulation2, based on their fitness.
    3.2 Update subpopulation1 by Lévy flight operator.
    3.3 Update subpopulation2 by flight straightly operator.
    3.4 GHS search
     If U(0, 1) ≤ 0.5
  Apply GHS algorithm on each individual X and generate the trial individual Y.
       Choose the best one of X and Y to enter the next generation.
    3.5 Baldwinian Learning
     If U(0, 1) ≤ 0.5
  Apply Baldwinin learning strategy on each individual X and generate the trial individual Y.
       Choose the best one of X and Y to enter the next generation.
    3.6 Apply transform function to obtain the potential solution of MKP.
    3.7 Repair the potential solution by Algorithm 3.
    3.8 Evaluate the fitness of the population and record the global best fitness.
G = G + 1.
    3.9 Sort the population by fitness.
   Step 4: End while
   Step 5: Output: the best results.
End.
The algorithm framework is shown in Figure 1.

3.5. Computational Complexity of One Iteration of HLMS

The computational complexity of one iteration of HLMS based on Algorithm 4 is described as follows.
(1)
The initialization of HLMS requires O(NP × n) time, where NP denotes the population size, and n is the dimension of MKP (the number of the items).
(2)
The discretization process of NP moth individual costs O(NP × n) time.
(3)
The quick repair operator takes O(n × m) time, where m is the constraints of the MKP instance.
(4)
Fitness evaluation has O(NP) time.
(5)
Lévy flight operator has O(NP1 × n) time, where NP1 is the number of individuals of subpopulation1.
(6)
Flight straightly operator has O(NP2 × n) time, where NP2 is the number of individuals of subpopulation2.
(7)
GHS learning requires O(NP × n) time.
(8)
Baldwinian learning requires O(NP × n) time.
(9)
Sort the population based on Quick sort algorithm and it takes time O(NPlogNP).
In summary, the total computational complexity is O(NP × n) per generation for fixed m.

4. Experimental Studies

To comprehensively evaluate the performance of the proposed HLMS, large numbers of experiments are implemented on the benchmark instances commonly used in the literature and a comparative study is conducted between HLMS and several populations based on optimization algorithms.

4.1. Benchmark Test Functions

In this paper, four sets of well-known benchmark data for MKP are used to test the performance of HLMS. These instances are described in [74] and available at OR-Library (http://people.brunel.ac.uk/~mastjjb/jeb/orlib/mknapinfo.html, accessed on 1 July 2021).
Test set I contains 18 small-scale instances with m = 2 to 30 and n = 20 to 105, which are denoted as Sento [76], HP [77], PB [77], and Weing [4].
Test set II contains 30 medium-scale instances with m = 5 and n = 30 to 90, which are marked as Weish [7].
Test set III contains 30 large-scale instances. These instances are divided into two subsets. Subset I includes 15 instances with m = 15 and n  {100, 250, 500}, which are labeled as cb1, cb2, and cb3, respectively. Subset II also includes 15 instances with m = 15 and n  {100, 250, 500}, which are called as cb4, cb5, and cb6, respectively.
Test set IV contains 9 instances with m {15, 25, 50} and n  {100, 200, 500, 1000, 1500}, which were created by Glover and Kochenberger and then are marked as GK.

4.2. Experimental Environment and Parameters Setting

The proposed HLMS algorithm includes several important parameters, whose values are empirically set based on the preliminary experiments and the details are recorded in Table 1.
The HLMS algorithm was implemented in C language and compiled using the GNU GCC compiler. All the experiments were carried out on a computer with Intel (R) Core (TM) i7-7500 CPU (2.90 GHz and 8.00 GB RAM), running the Windows 10 operating system. HLMS was independently run 30 times for each instance to eliminate the unfairness brought by the stochastic characteristic.
To comprehensively evaluate the performance of the HLMS algorithm, fourteen MKP algorithms in the literature are selected as our comparative algorithms, which are listed as follows.
  • Modified binary particle swarm optimization (MBPSO) [78];
  • Chaotic binary particle swarm optimization with time-varying acceleration coefficients (CBPSOTVAC) [79];
  • Binary PSO with time-varying acceleration coefficients (BPSOTVAC) [79];
  • Modified multi-verse optimization (MMVO) algorithm [17];
  • New binary particle swarm optimization with immunity-clonal algorithm (NPSOCLA) [80];
  • Binary gravitational search algorithm (BGSA) [81];
  • Binary hybrid topology particle swarm optimization (BHTPSO) [81];
  • Binary hybrid topology particle swarm optimization quadratic interpolation (BHTPSO-QI) [81];
  • New binary particle swarm optimization (NBPSO) [81];
  • Binary version of PSO (BPSO) [82];
  • Binary version of the Harris hawks algorithm (BHHA) [22,83];
  • Binary version of the salp swarm algorithm (BSSA) [84];
  • Binary version of the modified whale optimization algorithm (BIWOA) [85];
  • Binary version of the sin-cosine algorithm (BSCA) [86].
It should be noted that the results of the comparative algorithms are compiled from the related papers. If the result of an algorithm for a MKP instance is not available, the result of the instance is ignored. In addition, considering that different comparative algorithms are written in different programming languages, or run on different computing platforms based on different termination conditions and algorithm parameters, we focus on comparing solution quality.
For this purpose, in this paper, eight typical statistical evaluation criteria are selected to evaluate the performance of all the comparative algorithms.
  • Best value (Best):
B e s t = max ( f i ) , for i [ 1 , t ]
where  f i is the fitness value for  i t h time. t is the total number of independent experiments.
  • Worst value (Worst):
W o r s t = min ( f i ) , for i [ 1 , t ]
  • Mean value (Mean):
M e a n = 1 t i = 1 t f i
The mean value characterizes the centralized trend of the values of random variables. The larger the mean value, the more concentrated the results of multiple runs of the algorithm will be.
  • Standard deviation (Std):
S t d = 1 t i = 1 t ( f i m e a n )
Standard deviation describes the degree of dispersion of random variable values relative to the mean value. Meanwhile, standard deviation reflects the fluctuation in the value of a random variable. In other words, stability is an important evaluation criterion of a stochastic algorithm. If the Std value is high, the stability of the algorithm is poor, otherwise, its performance is good.
  • Success rate (SR):
S R = s t t
where st denotes the success times, that is, the number of the known theoretical optimal solution is obtained. The high success rate indicates that the algorithm has good stability and optimization performance.
  • Percent deviation (PDev):
P D e v = O p t M e a n O p t 100
where Opt represents the optimal or the best-known solution. PDev reflects the degree to which the mean value deviates from the known theoretical optimal solution when the algorithm solves a single instance.
  • Average error (AE):
A E = 1 N i = 1 N ( O p t p r o f i t ) O p t 100
Average error is an indicator that reflects the general level of error between random variables and Opt. Here, profit can be Best, Worst, or Mean. N is the number of benchmark instances. Clearly, the smaller AE value indicates that the algorithm has better performance. AE indicates the overall performance of the algorithm for solving a set of MKP instances.
  • Percentage gap (Gap):
G a p = O p t B e s t O p t 100
Similar to AE, for the maximization problem, the smaller the Gap is, the better the performance of the algorithm is. Gap investigates the performance of the algorithm to solve a single MKP instance.
Moreover, to determine whether there are significance differences between HLMS and other algorithms, the p-value based on the nonparametric Wilcoxon signed ranks test at the 95% confidence level is reported as well. Note that a p-value less than 0.05 represents that there exists a significant difference between the paired compared results. All the statistical results have been performed by the statistical software R language.

4.3. Comparisons on the Small-Scale Test Set

The proposed HLMS is first substantiated based on the 18 small-scale instances of Test set I and the experimental result is listed in Table 2, along with the available results of the comparative algorithms. In Table 2, the first three columns record the names of instances, the dimension information (n is the number of items, and m is the number of knapsacks), and the known optimum results (Opt), respectively. In addition, aggregate data are recorded at the bottom of the table. #Opt shows the number of the best-known solution obtained by the corresponding algorithm. #SR and #Std represent the number of instances for which the corresponding algorithm obtained a better result in terms of SR and Std among the comparative algorithms. MSR is the mean value of success rate and the ranks in descending order on the MSR are provided. Besides, the best findings among the comparison results are indicated in bold.
From Table 2, the proposed HLMS is able to obtain the known optimum solution for almost all the 18 instances except for Weing7. However, MS can only reach the known optimum solution for 13 instances. Considering the #SR, HLMS performs much better than the competing algorithms. In terms of MSR, HLMS is slightly worse than CBPSOTVAC and ranks two. Moreover, the clear superiority of HLMS is established in comparison with MS in terms of all evaluation criteria. Therefore, we can conclude that it is beneficial to use the GHS global search algorithm combined with the Baldwinian learning strategy. In terms of the p-value of SR, the difference between HLMS and MS, HLMS and MBPSO is statistically significant (p-value < 0.05). However, there are no significant differences in Std among the latter two groups (p-value > 0.05).
The related box plots are given in Figure 2 in terms of SR. As can be seen from Figure 2, the difference of SR among four algorithms is very obvious. Although the distributions of the SR value of MS and MBPSO are more uniform than that of HLMS and CBPSOTVAC, the interquartile ranges of the former are worse than that of the latter. Moreover, outliers exist in MS and MBPSO on the SR value. In addition, we also observe the maximum, upper quartile, the mean value of HLMS is equal to or close to 1.0, 0.8, and 0.6, respectively.
Based on the above analysis, we can draw a conclusion that HLMS can obtain the best-known solution of 18 small-scale instances with a high success rate.

4.4. Comparisons on the Medium-Scale Test Set

In the second experiment, HLMS is used to solve medium-scale test instance (Test set II) to verify the performance of algorithms. The results are reported in Table 3, together with the results of other six state-of-the-art MKP algorithms, including MS, BIWOA, BMMVO, BSCA, BHHA, and BSSA. Note that these six algorithms are all novel swarm intelligence algorithms proposed in recent years and it is meaningful to select them for comparative study of MKP.
From Table 3, the results demonstrate that HLMS reaches the optimum solutions on 27 out of 30 instances, while the six comparative algorithms obtain the optimum solutions only on 20, 16, 14, 18, 19, and 19, respectively. In terms of #Worst, BSSA outperforms the other six algorithms on 23 instances. PDev measures the deviation between the mean and the best-known solution. The small mean value of PDev also confirms the superiority of HLMS. Comprehensively speaking, MS has the worst performance among all the comparative algorithms. Moreover, there are significant differences (p-value < 0.05) between the comparisons of the first four groups concerning PDev.
Figure 3 presents the box plots of the PDev values for all the comparative algorithms. The span of each box implicitly reflects the stability of the algorithm. The smaller the span is, the better the stability of the algorithm is. As can be seen from Figure 3, HLMS has a significant advantage over the other six algorithms since the span of the box for HLMS is obviously smaller than that of the other comparative algorithms. It should be noted that the dot in Figure 3. represent outliers.
In summary, the results in Table 3 and Figure 3 indicate that our HLMS algorithm is very competitive compared to the other six MKP algorithms. The findings are based on the fact that the GHS learning scheme enhances the global search ability of HLMS. On this basis, Baldwinian learning can effectively adjust the shape of search space and thereby provides good search paths towards the best solutions.

4.5. Comparisons on the Large-Scale Test Set

In the third experiment, the performance of HLMS is verified by solving large-scale problems, and the comparative results on Test set III and Test set IV are reported in Table 4 and Table 5. In order to make a fair comparison with different classical algorithms using appropriate evaluation criteria, the experiment is divided into three groups on different scale instances.

4.5.1. Performance Comparison on Test Set III (cb1–cb3)

Table 4 summarizes the experimental results of the first group large-scale benchmarks. From Table 4, it can be seen clearly that the proposed HLMS still keeps the best performance in terms of all six evaluation criteria. Specifically speaking, HLMS outperforms the other comparative algorithms. In addition, the p-value indicates that there is significant difference between HLMS and MS, BGSA, BHTPSO, and BHTPSO-QI in terms of mean. However, the p-value is 0.201 for MMVO (p-value > 0.05) and then reject the null hypothesis. Hence, insignificant difference can be detected between HLMS and MMVO.
The performance comparison of six methods based on AE is plotted in Figure 4. It is evident that the axis of HLMS on radar charts has a point nearer to the center in comparison with the other five algorithms when considering AE of Best, AE of Mean, and AE of Worst, which indicates that it is more effective with respect to quality of solutions. It can be considered that HLMS obtained optimal or near-optimal for most of the instances in terms of Best, Mean, and Worst, and could beat all the other competing algorithms.
To observe the stability of MS and HLMS more intuitively, the error bar based on variance and the trends plot based on Std are shown in Figure 5 and Figure 6, respectively. It can be observed clearly from Figure 5 that the variances of HLMS are apparently smaller than that of MS for all benchmarks. Moreover, the variances will increase with the growth of the scale of instances. It is clear from Figure 6 that the trend lines of HLMS are located in the relatively low area for ten instances of cb1 and cb2. However, the curve of cb3 has an upward trend. In brief, the curve of MS is higher than that of HLMS, which indicates that HLMS has more stable performance than MS.

4.5.2. Performance Comparison on Test Set III (cb4–cb6)

Table 5 summarizes the experimental results of the second group large-scale benchmarks. Table 5 shows that HLMS is also very efficient for 15 large instances with m = 10. Moreover, HLMS is superior to other five algorithms in absolute advantage, which is confirmed by the small p-values (0.000 ≤ 0.05).
Similarly, radar charts are plotted to visualize three evaluation criteria, AE of Best, AE of Mean, and AE of Worst in Figure 7. From Figure 7, the phenomenon is almost consistent with Figure 4. The point on the HLMS axis is very close to the center point. By implication, HLMS has smaller AE value compared with other algorithms.
The error bar based on variance for HLMS and MS is illustrated in Figure 8, which is to assess the stability of algorithms. As can be seen from Figure 8, the variance of HLMS is almost unaffected by the scale of MKP. However, with the expansion of the scale, the variance of MS is increasing gradually.
The trend plot of Std for HLMS and MS is given in Figure 9. It can be seen from Figure 9 that the trend curve of MS is significantly higher than that of HLMS, which further indicates that HLMS has better stability than MS.

4.5.3. Performance Comparison on Test Set IV

Table 6 summarizes the experimental results of the third group large-scale benchmarks. Overall, HLMS still outperforms all other comparative algorithms. In terms of #Best, #Worst, and #Mean, HLMS obtains a better result respectively on 5, 1, and 3 out of 9 instances. The results of BIWOA are respectively on 2, 2, and 2 out of 9 instances. BSSA with better performance obtains 2, 6, and 5 out of 9 instances, respectively. For the significance, the p-values for BMMVO and BSCA are both smaller than 0.05 concerned with Mean, except for MS, BIWOA, BHHA, and BSSA, which indicates that the difference between HLMS and most comparative algorithms is not significant when facing Test set IV.
Graphically, Figure 10 shows the AE of Best, AE of Mean, and AE of Worst obtained by seven methods. In terms of Best, HLMS outperforms the other comparative algorithms in absolute small AE value. However, HLMS is slightly worse than BIWOA in terms of AE of Mean.
In summary, the above experimental results and comparisons show that the proposed HLMS also has excellent optimization performance in solving large-scale MKP instances, in terms of solution quality, convergence, and stability of algorithms. This is mainly due to GHS learning and Baldwinian learning being able to effectively balance exploration and exploitation in the evolution process. Compared with the original MS, the hybrid learning strategy focuses more on discovering and utilizing useful information from the whole population and whole search experience, rather than the experience of some random local individuals.

4.6. Sensitivity Analysis on the Positional Parameter and the Scale Parameter

As is known to all, Gaussian distribution and Cauchy distribution are two important distributions, which have been integrated into many algorithms and proved an effective strategy to enhance the ability of an elaborate search. The strength of Baldwinian learning in [47] is random real number obeying Gaussian distribution. However, previous studies have revealed that Cauchy mutation possesses more power in escaping local optima and converging to the global optimum. Hence, the parameters used in [47] for Gaussian distribution and the positional parameter x and the scale parameter y for Cauchy distribution are investigated in this subsection. It is noted that, to eliminate the influence of GHS, HLMS only adopted the Baldwinian learning strategy in this experiment. We tested HLMS for Gaussian distribution and different combinations for (x, y) of Cauchy distribution: (0, 1), (0, 0.5), (0, 2), and (−2, 1). Table 7 summaries the Mean and SR for 18 instances of Test set I. In addition, boxplot for five parameter combinations on SR is plotted in Figure 11.
From Table 7, all HLMS with four different combinations of x and y find better results than that of HLMS combined with Gaussian distribution. In terms of #Mean, #SR, and MSR, HLMS with four groups (x, y) shows similar results. It can be observed from Figure 11 that HLMS-C2 shows excellent comprehensive performance. HLMS-C2 has the best maximum and three-quarter quantile. Hence, considering all of the parameter combinations, we concluded that the setting x = 0 and y = 0.5 for the HLMS is an appropriate choice.
We can draw a conclusion from this experiment that it is better to use a random real obeying Cauchy distribution than Gaussian distribution as the Baldwinian learning strength. The reason may be that Cauchy mutation has stronger ability to jump from local optimum than Gaussian mutation.

4.7. The Effectiveness of the Two Components in HLMS

As mentioned above, HLMS includes two learning strategies: Baldwinian learning strategy and GHS learning strategy. The aim of this subsection is to investigate the effectiveness of these two learning strategies. Therefore, one additional experiment is conducted on Test set I and the results are summarized in Table 8. HLMS, which only adopted the Baldwinian learning strategy, is denoted as HLMS-B. HLMS, which only adopted the GHS learning strategy, is denoted as HLMS-H.
As can be seen from Table 8, compared with HLMS-B, HLMS-H, and HLMS, MS shows the worst performance in terms of #Mean, #SR, and MSR. The results further reveal that two learning strategies are effective in the search process. Additionally, it is noted that HLMS-B shows similar performance with MS while the performance difference between HLMS-H and MS is significant. However, the performance of HLMS integrated with two learning strategies is obviously better than that of any one.
Moreover, the convergence graphs of the average objective function values obtained by four algorithms are plotted in Figure 12 and Figure 13 for four representative instances: Sento1, Sento2, Weing7, and Weing8. As seen from Figure 12 and Figure 13, MS has the slowest convergence speed, while HLMS-H and HLMS have similar convergence speed and converge both faster than MS and HLMS-B.
In summary, we can draw a conclusion that the above two learning strategies can realize complementary advantages to enhance the performance of MS. Indeed, only using Baldwinian learning is not sufficient for exploitation. The conclusion is that the beneficial combination of the two strategies is significant for improving the performance of the algorithm.

4.8. Discussion

Comparison results demonstrate that the Baldwinian learning and GHS learning strategies can really improve the performance of HLMS, thereby making it better than the original MS and most other comparative algorithms on the majority of MKP instances, in terms of solution accuracy, convergence speed, and algorithm stability. The advantage of HLMS mainly lies in that GHS is used to guide the global search and dimensional learning can achieve a large jump to help solutions escape local extremum. The other reason is that Baldwinian learning as a local search strategy has the effect of changing the fitness landscape. This interaction between learning and evolution is very beneficial. Accordingly, both Baldwinian learning and GHS learning are more efficient and effective than MS alone.
In fact, based on the previous experimental results, we can find that considering small-scale MKP instances, medium-sized MKP instances, and large-scale MKP instances, the HLMS algorithm combining Baldwinian learning and GHS learning is an effective algorithm for solving MKP problems. Besides that, balancing exploration and exploitation is an important factor in metaheuristic algorithms by maintaining adequate diversity in swarm individuals so that reducing the probability of trapping in local optimal locations. In HLMS, GHS learning and Baldwinian learning respectively play the roles of exploration and exploitation, making the optimization performance of the algorithm better.

5. Conclusions and Future Work

This paper proposed a hybrid learning moth search algorithm (HLMS) inspired by the idea that the learning strategy could direct the evolutionary process. The framework proposed in this work includes two learning strategies: Baldwinian learning and GHS learning. In the search process, the two learning strategies play the role of local exploitation and global exploration, respectively.
HLMS is verified by solving the NP-hard 0–1 multidimensional knapsack problems. The experimental results on the 87 instances commonly used in literature showed that HLMS performs competitively in comparison with MS and other state-of-the-art meta-heuristics algorithms. Sensitivity analysis of Gaussian distribution and Cauchy distribution on Baldwinian learning is provided. The results proved that Cauchy mutation is more effective than Gaussian mutation as learning length. The effectiveness of two important learning strategies of HLMS is investigated. The results demonstrated that Baldwinian learning and GHS learning both play a major role in improving the performance of HLMS. MS with two learning strategies surpasses MS and MS with a single strategy. It confirms the effectiveness of our proposed learning strategies.
Future research on MS can be divided into two main directions: research on more real-world applications and research on improvements of the algorithms. Concerning the application of MS, it has the potential to solve more combinatorial optimization problems, such as maximum diversity problems (MDP), multi-objective knapsack problems (MOKP), and multi-demand multidimensional knapsack problems (MDMKP). In terms of the improvements of algorithms, more effective learning based on strategies can be adopted to enhance the search ability of the algorithm, such as orthogonal learning, reinforcement learning, and adaptive learning.

Author Contributions

Y.F. performed the methodology, investigation and wrote the draft. H.W. and Z.C. supervised the research and edited the final draft. X.L. performed the experiments. M.L. reviewed the final draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (No. 61806069), Key projects of science and technology research in Colleges of Hebei Province, (No. ZD2022083), Key R & D plan project of Hebei Province (No. 22375415D), Major Projects of Guangdong Education Department for Foundation Research and Applied Research (No. 2017KZDXM081, 2018KZDXM066), Guangdong Provincial University Innovation Team Project (No. 2020KCXTD045).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Toyoda, Y. A simplified algorithm for obtaining approximate solutions to zero-one programming problems. Manag. Sci. 1975, 21, 1417–1427. [Google Scholar] [CrossRef]
  2. Vasquez, M.; Vimont, Y. Improved results on the 0–1 multidimensional knapsack problem. Eur. J. Oper. Res. 2005, 165, 70–81. [Google Scholar] [CrossRef]
  3. Feng, Y.; Yang, J.; Wu, C.; Lu, M.; Zhao, X.-J. Solving 0-1 knapsack problems by chaotic monarch butterfly optimization algorithm with Gaussian mutation. Memetic Comput. 2018, 10, 135–150. [Google Scholar] [CrossRef]
  4. Weingartner, H.M.; Ness, D.N. Methods for the solution of the multidimensional 0/1 knapsack problem. Oper. Res. 1967, 15, 83–103. [Google Scholar] [CrossRef]
  5. Drake, J.H.; Özcan, E.; Burke, E.K. A case study of controlling crossover in a selection hyper-heuristic framework using the multidimensional knapsack problem. Evol. Comput. 2016, 24, 113–141. [Google Scholar] [CrossRef] [Green Version]
  6. Gilmore, P.; Gomory, R.E. The theory and computation of knapsack functions. Oper. Res. 1966, 14, 1045–1074. [Google Scholar] [CrossRef]
  7. Shih, W. A branch and bound method for the multiconstraint zero-one knapsack problem. J. Oper. Res. Soc. 1979, 30, 369–378. [Google Scholar] [CrossRef]
  8. Boussier, S.; Vasquez, M.; Vimont, Y.; Hanafi, S.; Michelon, P. A multi-level search strategy for the 0-1 Multidimensional Knapsack Problem. Discret. Appl. Math. 2010, 158, 97–109. [Google Scholar] [CrossRef]
  9. Lai, X.; Hao, J.-K.; Glover, F.; Lü, Z. A two-phase tabu-evolutionary algorithm for the 0-1 multidimensional knapsack problem. Inf. Sci. 2018, 436, 282–301. [Google Scholar] [CrossRef]
  10. Chih, M. Three pseudo-utility ratio-inspired particle swarm optimization with local search for multidimensional knapsack problem. Swarm Evol. Comput. 2018, 39, 279–296. [Google Scholar] [CrossRef]
  11. Haddar, B.; Khemakhem, M.; Hanafi, S.; Wilbaut, C. A hybrid quantum particle swarm optimization for the multidimensional knapsack problem. Eng. Appl. Artif. Intell. 2016, 55, 1–13. [Google Scholar] [CrossRef]
  12. Lai, X.; Hao, J.-K.; Fu, Z.-H.; Yue, D. Diversity-preserving quantum particle swarm optimization for the multidimensional knapsack problem. Expert Syst. Appl. 2020, 149, 113310. [Google Scholar] [CrossRef]
  13. Wang, L.; Wang, S.-y.; Xu, Y. An effective hybrid EDA-based algorithm for solving multidimensional knapsack problem. Expert Syst. Appl. 2012, 39, 5593–5599. [Google Scholar] [CrossRef]
  14. Li, Z.; Tang, L.; Liu, J. A memetic algorithm based on probability learning for solving the multidimensional knapsack problem. IEEE Trans. Cybern. 2020, 54, 2284–2299. [Google Scholar] [CrossRef] [PubMed]
  15. Luo, K.; Zhao, Q. A binary grey wolf optimizer for the multidimensional knapsack problem. Appl. Soft Comput. 2019, 83, 105645. [Google Scholar] [CrossRef]
  16. Zhang, B.; Pan, Q.K.; Zhang, X.L.; Duan, P.Y. An effective hybrid harmony search-based algorithm for solving multidimensional knapsack problems. Appl. Soft Comput. 2015, 29, 288–297. [Google Scholar] [CrossRef]
  17. Abdel-Basset, M.; El-Shahat, D.; Faris, H.; Mirjalili, S. A binary multi-verse optimizer for 0-1 multidimensional knapsack problems with application in interactive multimedia systems. Comput. Ind. Eng. 2019, 132, 187–206. [Google Scholar] [CrossRef]
  18. García, J.; Maureira, C. A KNN quantum cuckoo search algorithm applied to the multidimensional knapsack problem. Appl. Soft Comput. 2021, 102, 107077. [Google Scholar] [CrossRef]
  19. Bolaji, A.L.a.; Okwonu, F.Z.; Shola, P.B.; Balogun, B.S.; Adubisi, O.D. A modified binary pigeon-inspired algorithm for solving the multi-dimensional knapsack problem. J. Intell. Syst. 2021, 30, 90–103. [Google Scholar] [CrossRef]
  20. Feng, Y.; Wang, G.-G. A binary moth search algorithm based on self-learning for multidimensional knapsack problems. Future Gener. Comput. Syst. 2022, 126, 48–64. [Google Scholar] [CrossRef]
  21. Gupta, S.; Su, R.; Singh, S. Diversified sine-cosine algorithm based on differential evolution for multidimensional knapsack problem. Appl. Soft Comput. 2022, 130, 109682. [Google Scholar] [CrossRef]
  22. Abdel-Basset, M.; Mohamed, R.; Sallam, K.M.; Chakrabortty, R.K.; Ryan, M.J. BSMA: A novel metaheuristic algorithm for Multi-dimensional knapsack problems: Method and comprehensive analysis. Comput. Ind. Eng. 2021, 159, 107469. [Google Scholar] [CrossRef]
  23. Li, W.; Yao, X.; Zhang, T.; Wang, R.; Wang, L. Hierarchy ranking method for multimodal multi-objective optimization with local pareto fronts. IEEE Trans. Evol. Comput. 2022, 27, 98–110. [Google Scholar] [CrossRef]
  24. Cacchiani, V.; Iori, M.; Locatelli, A.; Martello, S. Knapsack problems-An overview of recent advances. Part II: Multiple, multidimensional, and quadratic knapsack problems. Comput. Oper. Res. 2022, 143, 105693. [Google Scholar] [CrossRef]
  25. Gao, D.; Wang, G.-G.; Pedrycz, W. Solving fuzzy job-shop scheduling problem using DE algorithm improved by a selection mechanism. IEEE Trans. Fuzzy Syst. 2020, 28, 3265–3275. [Google Scholar] [CrossRef]
  26. Wang, G.-G.; Gao, D.; Pedrycz, W. Solving multi-objective fuzzy job-shop scheduling problem by a hybrid adaptive differential evolution algorithm. IEEE Trans. Ind. Inform. 2022, 18, 8519–8528. [Google Scholar] [CrossRef]
  27. Shadkam, E.; Bijari, M. A novel improved cuckoo optimisation algorithm for engineering optimisation. Int. J. Artif. Intell. Soft Comput. 2020, 7, 164–177. [Google Scholar] [CrossRef]
  28. Parashar, S.; Senthilnath, J.; Yang, X.-S. A novel bat algorithm fuzzy classifier approach for classification problems. Int. J. Artif. Intell. Soft Comput. 2017, 6, 108–128. [Google Scholar] [CrossRef]
  29. Wang, G.-G.; Deb, S.; Gandomi, A.H.; Alavi, A.H. Opposition-based krill herd algorithm with cauchy mutation and position clamping. Neurocomputing 2016, 177, 147–157. [Google Scholar] [CrossRef]
  30. Wang, G.-G.; Guo, L.; Gandomi, A.H.; Hao, G.S.; Wang, H. Chaotic krill herd algorithm. Inf. Sci. 2014, 274, 17–34. [Google Scholar] [CrossRef]
  31. Li, W.; Wang, G.-G.; Alavi, A.H. Learning-based elephant herding optimization algorithm for solving numerical optimization problems. Knowl.-Based Syst. 2020, 195, 105675. [Google Scholar] [CrossRef]
  32. Wang, G.-G.; Deb, S.; Cui, Z. Monarch butterfly optimization. Neural Comput. Appl. 2019, 31, 1995–2014. [Google Scholar] [CrossRef] [Green Version]
  33. Feng, Y.; Deb, S.; Wang, G.-G.; Alavi, A.H. Monarch butterfly optimization: A comprehensive review. Expert Syst. Appl. 2021, 168, 114418. [Google Scholar] [CrossRef]
  34. Ma, L.; Cheng, S.; Shi, Y. Enhancing learning efficiency of brain storm optimization via orthogonal learning design. IEEE Trans. Syst. Man Cybern. Syst. 2020, 51, 6723–6742. [Google Scholar] [CrossRef]
  35. Yu, H.; Li, W.; Chen, C.; Liang, J.; Gui, W.; Wang, M.; Chen, H. Dynamic Gaussian bare-bones fruit fly optimizers with abandonment mechanism: Method and analysis. Eng. Comput. 2020, 38, 743–771. [Google Scholar] [CrossRef]
  36. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  37. Zhou, Y.; Bao, Z.; Luo, Q.; Zhang, S. A complex-valued encoding wind driven optimization for the 0-1 knapsack problem. Appl. Intell. 2017, 46, 684–702. [Google Scholar] [CrossRef]
  38. Song, J.; Chen, C.; Heidari, A.A.; Liu, J.; Yu, H.; Chen, H. Performance optimization of annealing salp swarm algorithm: Frameworks and applications for engineering design. J. Comput. Des. Eng. 2022, 9, 633–669. [Google Scholar] [CrossRef]
  39. Gezici, H.; Livatyalı, H. Chaotic Harris hawks optimization algorithm. J. Comput. Des. Eng. 2022, 9, 216–245. [Google Scholar] [CrossRef]
  40. Chou, J.S.; Truong, D.N. A novel metaheuristic optimizer inspired by behavior of jellyfish in ocean. Appl. Math. Comput. 2021, 389, 125535. [Google Scholar] [CrossRef]
  41. Wang, L.Y.; Cao, Q.J.; Zhang, Z.X.; Mirjalili, S.; Zhao, W.G. Artificial rabbits optimization: A new bio-inspired meta-heuristic algorithm for solving engineering optimization problems. Eng. Appl. Artif. Intell. 2022, 114, 31. [Google Scholar] [CrossRef]
  42. Abdollahzadeh, B.; Gharehchopogh, F.S.; Khodadadi, N.; Mirjalili, S. Mountain Gazelle Optimizer: A new Nature-inspired Metaheuristic Algorithm for Global Optimization Problems. Adv. Eng. Softw. 2022, 174, 34. [Google Scholar] [CrossRef]
  43. Chen, C.; Wang, X.; Yu, H.; Wang, M.; Chen, H. Dealing with multi-modality using synthesis of Moth-flame optimizer with sine cosine mechanisms. Math. Comput. Simul. 2021, 188, 291–318. [Google Scholar] [CrossRef]
  44. Wang, G.-G. Moth search algorithm: A bio-inspired metaheuristic algorithm for global optimization problems. Memetic Comput. 2018, 10, 151–164. [Google Scholar] [CrossRef]
  45. Li, J.; Yang, Y.H.; An, Q.; Lei, H.; Deng, Q.; Wang, G.G. Moth Search: Variants, Hybrids, and Applications. Mathematics 2022, 10, 4162. [Google Scholar] [CrossRef]
  46. Elaziz, M.A.; Xiong, S.; Jayasena, K.; Li, L. Task scheduling in cloud computing based on hybrid moth search algorithm and differential evolution. Knowl.-Based Syst. 2019, 169, 39–52. [Google Scholar] [CrossRef]
  47. Strumberger, I.; Sarac, M.; Markovic, D.; Bacanin, N. Moth search algorithm for drone placement problem. Int. J. Comput. 2018, 3, 75–80. [Google Scholar]
  48. Strumberger, I.; Tuba, E.; Bacanin, N.; Beko, M.; Tuba, M. Hybridized moth search algorithm for constrained optimization problems. In Proceedings of the 2018 International Young Engineers Forum (YEF-ECE), Monte de Caparica, Portugal, 4 May 2018; pp. 1–5. [Google Scholar]
  49. Feng, Y.; Wang, G.-G. Binary moth search algorithm for discounted {0-1} knapsack problem. IEEE Access 2018, 6, 10708–10719. [Google Scholar] [CrossRef]
  50. Feng, Y.; An, H.; Gao, X. The importance of transfer function in solving set-union knapsack problem based on discrete moth search algorithm. Mathematics 2019, 7, 17. [Google Scholar] [CrossRef] [Green Version]
  51. Feng, Y.; Yi, J.-H.; Wang, G.-G. Enhanced moth search algorithm for the set-union knapsack problems. IEEE Access 2019, 7, 173774–173785. [Google Scholar] [CrossRef]
  52. Cui, Z.; Xue, F.; Cai, X.; Cao, Y.; Wang, G.-G.; Chen, J. Detection of malicious code variants based on deep learning. IEEE Trans. Ind. Inform. 2018, 14, 3187–3196. [Google Scholar] [CrossRef]
  53. Zhang, J.; Pan, L.; Han, Q.-L.; Chen, C.; Wen, S.; Xiang, Y. Deep Learning Based Attack Detection for Cyber-Physical System Cybersecurity: A Survey. IEEE/CAA J. Autom. Sin. 2021, 9, 377–391. [Google Scholar] [CrossRef]
  54. Wang, F.; Wang, X.; Sun, S. A reinforcement learning level-based particle swarm optimization algorithm for large-scale optimization. Inf. Sci. 2022, 602, 298–312. [Google Scholar] [CrossRef]
  55. Wang, L.; Pan, Z.; Wang, J. A review of reinforcement learning based intelligent optimization for manufacturing scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  56. Zhou, T.; Chen, M.; Zou, J. Reinforcement learning based data fusion method for multi-sensors. IEEE/CAA J. Autom. Sin. 2020, 7, 1489–1497. [Google Scholar] [CrossRef]
  57. Zheng, R.-z.; Zhang, Y.; Yang, K. A transfer learning-based particle swarm optimization algorithm for travelling salesman problem. J. Comput. Des. Eng. 2022, 9, 933–948. [Google Scholar] [CrossRef]
  58. Bingjie, X.; Deming, L. Q-Learning-based teaching-learning optimization for distributed two-stage hybrid flow shop scheduling with fuzzy processing time. Complex Syst. Model. Simul. 2022, 2, 113–129. [Google Scholar] [CrossRef]
  59. Wang, G.-G.; Tan, Y. Improving metaheuristic algorithms with information feedback models. IEEE Trans. Cybern. 2019, 49, 542–555. [Google Scholar] [CrossRef]
  60. Chen, C.; Wang, X.; Yu, H.; Zhao, N.; Wang, M.; Chen, H.; Kumarappan, N. An Enhanced Comprehensive Learning Particle Swarm Optimizer with the Elite-Based Dominance Scheme. Complexity 2020, 2020, 1–24. [Google Scholar] [CrossRef]
  61. Gong, M.; Jiao, L.; Zhang, L. Baldwinian learning in clonal selection algorithm for optimization. Inf. Sci. 2010, 180, 1218–1236. [Google Scholar] [CrossRef]
  62. Hinton, G.E.; Nowlan, S.J. How learning can guide evolution. In Adaptive Individuals in Evolving Populations: Models and Algorithms; Addison-Wesley Longman Publishing Co., Inc.: Boston, MA, USA, 1996; Volume 26, pp. 447–454. [Google Scholar]
  63. Qi, Y.; Liu, F.; Liu, M.; Gong, M.; Jiao, L. Multi-objective immune algorithm with Baldwinian learning. Appl. Soft Comput. 2012, 12, 2654–2674. [Google Scholar] [CrossRef]
  64. Zhang, C.; Chen, J.; Xin, B. Distributed memetic differential evolution with the synergy of Lamarckian and Baldwinian learning. Appl. Soft Comput. 2013, 13, 2947–2959. [Google Scholar] [CrossRef]
  65. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A new heuristic optimization algorithm: Harmony search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  66. Omran, M.G.H.; Mahdavi, M. Global-best harmony search. Appl. Math. Comput. 2008, 198, 643–656. [Google Scholar] [CrossRef]
  67. Xiang, W.L.; An, M.Q.; Li, Y.Z.; He, R.C.; Zhang, J.F. A novel discrete global-best harmony search algorithm for solving 0-1 knapsack problems. Discret. Dyn. Nat. Soc. 2014, 2014, 1–12. [Google Scholar] [CrossRef] [Green Version]
  68. Keshtegar, B.; Sadeq, M.O. Gaussian global-best harmony search algorithm for optimization problems. Soft Comput. 2017, 21, 7337–7349. [Google Scholar] [CrossRef]
  69. El-Abd, M. An improved global-best harmony search algorithm. Appl. Math. Comput. 2013, 222, 94–106. [Google Scholar] [CrossRef]
  70. Peng, Y.; Lu, B.-L. Hybrid learning clonal selection algorithm. Inf. Sci. 2015, 296, 128–146. [Google Scholar] [CrossRef]
  71. Lan, K.-T.; Lan, C.-H. Notes on the distinction of Gaussian and Cauchy mutations. In Proceedings of the 2008 Eighth International Conference on Intelligent Systems Design and Applications, Kaohsuing, Taiwan, 26–28 November 2008; pp. 272–277. [Google Scholar]
  72. Mahdavi, M.; Fesanghary, M.; Damangir, E. An improved harmony search algorithm for solving optimization problems. Appl. Math. Comput. 2007, 188, 1567–1579. [Google Scholar] [CrossRef]
  73. Xiao, S.; Wang, W.; Wang, H.; Tan, D.; Wang, Y.; Yu, X.; Wu, R. An Improved Artificial Bee Colony Algorithm Based on Elite Strategy and Dimension Learning. Mathematics 2019, 7, 289. [Google Scholar] [CrossRef] [Green Version]
  74. Chu, P.C.; Beasley, J.E. A genetic algorithm for the multidimensional knapsack problem. J. Heuristics 1998, 4, 63–86. [Google Scholar] [CrossRef]
  75. Puchinger, J.; Raidl, G.R.; Pferschy, U. The multidimensional knapsack problem: Structure and algorithms. Inf. J. Comput. 2010, 22, 250–265. [Google Scholar] [CrossRef] [Green Version]
  76. Senju, S.; Toyoda, Y. An approach to linear programming with 0-1 variables. Manag. Sci. 1968, 15, B196–B207. [Google Scholar] [CrossRef]
  77. Freville, A. Hard 0-1 multiknapsack test problems for size reduction methods. Investig. Oper. 1990, 1, 251–270. [Google Scholar]
  78. Bansal, J.C.; Deep, K. A modified binary particle swarm optimization for knapsack problems. Appl. Math. Comput. 2012, 218, 11042–11061. [Google Scholar] [CrossRef]
  79. Chih, M.; Lin, C.-J.; Chern, M.-S.; Ou, T.-Y. Particle swarm optimization with time-varying acceleration coefficients for the multidimensional knapsack problem. Appl. Math. Model. 2014, 38, 1338–1350. [Google Scholar] [CrossRef]
  80. Dina, E.-G.; Badr, A.; Abd El Azeim, M. New binary particle swarm optimization with immunity-clonal algorithm. J. Comput. Sci. 2013, 9, 1534. [Google Scholar]
  81. Beheshti, Z.; Shamsuddin, S.M.; Hasan, S. Memetic binary particle swarm optimization for discrete optimization problems. Inf. Sci. 2015, 299, 58–84. [Google Scholar] [CrossRef]
  82. Kennedy, J.; Eberhart, R.C. A discrete binary version of the particle swarm algorithm. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics, Orlando, FL, USA, 12–15 October 1997; pp. 4104–4108. [Google Scholar]
  83. Aaha, B.; Sm, C.; Hf, D.; Ia, D.; Mm, E.; Hc, F. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar]
  84. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp swarm algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  85. Abdel-Basset, M.; El-Shahat, D.; Sangaiah, A.K. A modified nature inspired meta-heuristic whale optimization algorithm for solving 0-1 knapsack problem. Int. J. Mach. Learn. Cybern. 2017, 10, 1–20. [Google Scholar] [CrossRef]
  86. Pinto, H.; Pea, A.; Valenzuela, M.; Fernández, A. A binary sine-cosine algorithm applied to the knapsack problem. In Proceedings of the Computer Science Online Conference, Zlin, Czech Republic, 24–27 April 2019; pp. 128–138. [Google Scholar]
Figure 1. The framework of HLMS for MKP.
Figure 1. The framework of HLMS for MKP.
Mathematics 11 01811 g001
Figure 2. Boxplot for 4 comparative algorithms on SR for Test set I.
Figure 2. Boxplot for 4 comparative algorithms on SR for Test set I.
Mathematics 11 01811 g002
Figure 3. Boxplot for 7 comparative algorithms on PDev for Test set II.
Figure 3. Boxplot for 7 comparative algorithms on PDev for Test set II.
Mathematics 11 01811 g003
Figure 4. The performance comparison of 6 methods based on AE for Test set III (cb1–cb3).
Figure 4. The performance comparison of 6 methods based on AE for Test set III (cb1–cb3).
Mathematics 11 01811 g004
Figure 5. The error bars (Variance) for MS and HLMS on Test set III (cb1–cb3).
Figure 5. The error bars (Variance) for MS and HLMS on Test set III (cb1–cb3).
Mathematics 11 01811 g005
Figure 6. The trends of Std for MS and HLMS on Test set III (cb1–cb3).
Figure 6. The trends of Std for MS and HLMS on Test set III (cb1–cb3).
Mathematics 11 01811 g006
Figure 7. The performance comparison of 5 methods based on AE for Test set III (cb4–cb6).
Figure 7. The performance comparison of 5 methods based on AE for Test set III (cb4–cb6).
Mathematics 11 01811 g007
Figure 8. Error bars (Variance) for MS and HLMS on Test set III (cb4–cb6).
Figure 8. Error bars (Variance) for MS and HLMS on Test set III (cb4–cb6).
Mathematics 11 01811 g008
Figure 9. The trends of Std for MS and HLMS on Test set III (cb4–cb6).
Figure 9. The trends of Std for MS and HLMS on Test set III (cb4–cb6).
Mathematics 11 01811 g009
Figure 10. The performance comparison of 7 methods based on AE for Test set IV.
Figure 10. The performance comparison of 7 methods based on AE for Test set IV.
Mathematics 11 01811 g010
Figure 11. Boxplot for 5 parameter combinations on SR for Test set I.
Figure 11. Boxplot for 5 parameter combinations on SR for Test set I.
Mathematics 11 01811 g011
Figure 12. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Sento1 and Sento2.
Figure 12. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Sento1 and Sento2.
Mathematics 11 01811 g012
Figure 13. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Weing7 and Weing8.
Figure 13. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Weing7 and Weing8.
Mathematics 11 01811 g013
Table 1. Settings of parameters of HLMS.
Table 1. Settings of parameters of HLMS.
ParametersSectionDescriptionValues
Smax2.1The max step used in Equation (5)1.0
  φ 2.1The acceleration factor used in Equation (7)0.618
  λ 2.1The scale factor used in Equation (7)a random number of uniform distribution in [0, 1]
  β 2.1Parameter used in Equation (6)1.5
c2.2The strength of Baldwinian learningA random number based on Cauchy distribution
PARmax2.3Parameter used in Equation (9)0.99
PARmin2.3Parameter used in Equation (9)0.01
HMCR2.3Parameter used in Algorithm 20.9
NP3.4The population size50
NFE3.4The maximal number of function evaluation100,000
Table 2. The results of HLMS with 3 comparative algorithms for Test set I.
Table 2. The results of HLMS with 3 comparative algorithms for Test set I.
Prob.n × mOpt.MSHLMSMBPSOCBPSOTVAC
SRStdSRStdSRStdSRStd
Sento160 × 3077720.1754.460.9033.730.1643.230.39357.78
Sento260 × 3087220.0027.790.474.720.0318.800.20101.03
HP128 × 434180.1323.110.4019.960.1025.520.3810.69
HP235 × 431860.0026.470.4032.700.1139.150.5921.35
PB127 × 430900.0330.890.4317.160.1124.320.4010.52
PB234 × 431860.0019.950.7013.830.1639.310.5118.73
PB429 × 295,1680.00894.680.301521.730.2718030.84875.1
PB520 × 1021390.1723.880.7021.260.0824.360.806.83
PB640 × 307760.4324.270.8017.280.2829.120.5440.17
PB737 × 3010350.033.860.505.830.0516.290.4024.25
Weing128 × 214,12780.90214.940.9389.770.82250.430.92281.98
Weing228 × 2130,8830.305731.340.9729.210.65314.080.88545.50
Weing328 × 295,6770.133767.420.801996.820.11876.780.75672.42
Weing428 × 2119,3370.731329.640.37711.430.761270.800.97378.58
Weing528 × 298,7960.232671.331.000.000.521923.50.94572.82
Weing628 × 2130,6230.23164.950.9771.200.36322.400.97343.45
Weing7105 × 21,095,4450.00482.740.002872.290.021130.600.0030,020.00
Weing8105 × 2624,3190.331966.370.031135.870.034704.300.2075,169.00
#Opt13 17 18 17
#SR0 12 1 5
#Std1 10 0 7
MSR0.21 0.54 0.26 0.59
Rank of MSR4 2 3 1
p-value0.0010.029 0.0010.6630.6410.896
Table 3. The results of HLMS with 6 comparative algorithms for Test set II.
Table 3. The results of HLMS with 6 comparative algorithms for Test set II.
Prob.n × mOpt. MSHLMSBIWOABMMVOBSCABHHABSSA
Weish0130 × 54554Best4554455445544554455445544554
Worst4477453445544554453445544554
PDev0.380.180.000.000.090.000.00
Weish0230 × 54536Best4536453645364536453645364536
Worst4440450445364536453645364536
PDev0.380.020.000.000.000.000.00
Weish0330 × 54115Best4106411541064106410641064106
Worst4106410641064106410641064106
PDev0.220.193.973.050.002.243.97
Weish0430 × 54561Best4561456145614561456145614561
Worst4505453145614561456145614561
PDev0.370.090.000.000.000.000.00
Weish0530 × 54514Best4514451445144514451445144514
Worst4514451445144514451445144514
PDev0.000.000.000.000.000.000.00
Weish0640 × 55557Best5557555755575557555755575557
Worst5502554255425542554255445557
PDev0.400.090.120.140.140.020.00
Weish0740 × 55567Best5567556755675567556755675567
Worst5360554255675567556755675567
PDev0.530.030.000.000.000.000.00
Weish0840 × 55605Best5605560556055605560556055605
Worst5478560356055603560356055605
PDev0.300.010.000.030.010.000.00
Weish0940 × 55246Best5246524652465246524652465246
Worst5185524652465246524652465246
PDev0.080.000.000.000.000.000.00
Weish1050 × 56339Best6339633963236303630363036303
Worst6255628063036303630363036303
PDev0.550.040.310.560.560.560.56
Weish1150 × 55643Best5643564356435643564356435643
Worst5592564356435643564356435643
PDev0.030.000.000.000.000.000.00
Weish1250 × 56339Best6339633963026301630263026302
Worst6090630463026301630163016301
PDev0.940.070.580.590.590.590.59
Weish1350 × 56159Best6159615961596159615961596159
Worst6025602561596159615961596159
PDev0.980.230.000.000.000.000.00
Weish1460 × 56954Best6954695469236923692369236923
Worst6769690269006900690069236923
PDev0.800.280.710.570.660.440.44
Weish1560 × 57486Best7486748674867486748674867486
Worst7199744274537449748674867486
PDev0.840.080.080.110.000.000.00
Weish1660 × 57289Best7289728972897289728972897289
Worst6942722172887281728172817281
PDev0.960.140.010.100.090.100.10
Weish1760 × 58633Best8633863386338624863386338633
Worst8141863385758497850686338633
PDev0.310.000.390.960.420.000.00
Weish1870 × 59580Best9540958095609456957395809573
Worst8857952594619318945195219527
PDev1.420.110.651.920.620.270.17
Weish1970 × 57698Best7698769876987698769876987698
Worst7448767476327629769876987698
PDev0.850.010.380.350.000.000.00
Weish2070 × 59450Best9450945094509445945094509450
Worst9306940894009365943394459450
PDev0.490.030.230.570.020.010.00
Weish2170 × 59074Best9074907490749074907490749074
Worst8922900890168969903390749074
PDev0.440.050.190.640.030.000.00
Weish2280 × 58947Best8790892989098886890989128912
Worst7904870889088886888689098912
PDev5.510.580.430.680.630.390.39
Weish2380 × 58344Best8170834483038250834483448344
Worst7246815482458233824582508303
PDev5.740.850.661.160.890.710.26
Weish2480 × 510,220Best10,18910,22010,18910,05810,21510,20210,220
Worst980710,09110,053978710,04210,13410,132
PDev1.700.161.233.140.860.570.48
Weish2580 × 59939Best9922993998859844993999399939
Worst9703988598089710988599159915
PDev1.020.030.941.630.200.210.11
Weish2690 × 59584Best9581958495759575957595759575
Worst8904951494779439947694889575
PDev1.660.190.691.170.920.240.09
Weish2790 × 59819Best9764981997789589976497649764
Worst9319976497739487963196789764
PDev2.090.500.452.530.880.610.56
Weish2890 × 59492Best9492949294549400945494549454
Worst9034943894119183940094009400
PDev1.540.190.491.700.780.620.45
Weish2990 × 59410Best9369941093699369936993699369
Worst8927936993699135936993699369
PDev1.450.400.431.750.430.430.43
Weish3090 × 511,191Best11,14811,18711,12111,02511,16911,16911,169
Worst10,80811,15510,97910,79010,94811,13511,154
PDev1.440.081.232.490.610.270.20
#Opt 20271614181919
#Worst 291210121823
The mean of PDev 1.110.150.470.860.310.280.29
p-value (PDev) 0.000 0.0110.0000.0330.3280.848
Table 4. The results of HLMS with 5 comparative algorithms for Test set III (cb1-cb3).
Table 4. The results of HLMS with 5 comparative algorithms for Test set III (cb1-cb3).
Prob.n × mOpt.ProfitMSHLMSMMVOBGSABHTPSOBHTPSO-QI
cb1-1100 × 524,381Best24,25324,38124,19224,15224,16924,301
Mean24,00424,30124,05023,83523,82223,821
Worst23,31124,23823,92023,17523,41523,287
cb1-2100 × 524,274Best24,25824,27424,27423,98624,10923,944
Mean23,93424,23124,27423,53623,65723,688
Worst23,36624,11624,27423,17722,95323,375
cb1-3100 × 523,551Best23,53823,55123,53823,38623,43523,418
Mean23,27223,52123,52023,04123,07223,073
Worst22,95323,46823,49422,54322,67822,621
cb1-4100 × 523,534Best23,25623,50323,28823,17223,25323,192
Mean23,02423,42023,12022,86322,92822,923
Worst22,54223,28823,04222,46822,50722,234
cb1-5100 × 523,991Best23,84523,96623,94723,75523,81523,774
Mean23,56723,93723,90023,45923,47323,527
Worst23,06223,83623,79923,10623,15523,053
cb2-1250 × 559,312Best58,08459,06358,47357,56557,81457,800
Mean57,36958,86258,24056,55456,87456,685
Worst55,98458,65358,11255,19154,93555,255
cb2-2250 × 561,472Best60,24861,29560,69260,05759,98259,767
Mean59,38661,05160,39058,61358,58858,680
Worst58,16760,87060,19457,70756,80756,821
cb2-3250 × 562,130Best61,21261,76761,70259,93660,63060,524
Mean59,92261,55261,33058,97559,23459,186
Worst57,88561,30361,15857,72357,43557,278
cb2-4250 × 559,463Best58,38659,14058,44157,97057,73657,884
Mean57,75258,92258,30056,74456,77356,584
Worst56,76358,71058,16355,37155,58955,164
cb2-5250 × 558,951Best57,75558,60558,08256,95957,37857,550
Mean56,92958,39058,30055,96156,12956,361
Worst56,32658,08858,16354,63754,36453,929
cb3-1500 × 5120,148Best116,296119,101119,978111,206114,493114,438
Mean115,444118,457119,900108,930111,017111,469
Worst114,634117,842119,810106,951106,454107,005
cb3-2500 × 5117,879Best113,732116,227115,634108,522112,821112,147
Mean112,257115,704115,400106,631109,276109,247
Worst111,381115,053115,222104,519100,118104,696
cb3-3500 × 5121,131Best117,666119,990119,156111,271114,774116,099
Mean116,367119,468118,900109,430112,035112,001
Worst115,160119,054118,651107,683106,406104,627
cb3-4500 × 5120,804Best116,454119,015119,124111,283115,828114,327
Mean115,396118,386118,900109,062112,200111,671
Worst114,100117,572118,623107,061106,222107,578
cb3-5500 × 5122,319Best117,900120,918121,141112,391115,889117,242
Mean116,767120,278120,800110,564112,253113,364
Worst115,062119,519120,401108,670102,820103,910
#Best 0124000
#Mean 0114000
#Worst 087000
AE of Best 1.91%0.57%0.98%3.97%2.70%2.72%
AE of Mean 3.08%0.92%1.23%5.63%4.81%4.75%
AE of Worst 4.81%1.36%1.48%7.49%8.29%8.09%
Rank of AE of Best 312645
p-value (Mean) 0.000 0.2010.0000.0000.000
Table 5. The results of HLMS with 5 comparative algorithms for Test set III (cb4–cb5).
Table 5. The results of HLMS with 5 comparative algorithms for Test set III (cb4–cb5).
Prob.n × mOpt.ProfitMSHLMSBGSABHTPSOBHTPSO-QI
cb4-1100 × 1023,064Best22,75323,05522,83622,90522,876
Mean22,45922,91422,33422,42522,449
Worst22,08022,75321,97521,98021,999
cb4-2100 × 1022,801Best22,61122,74322,44122,57322,408
Mean22,25522,62921,99122,04722,017
Worst21,62222,40721,43521,32221,454
cb4-3100 × 1022,131Best21,88622,13121,84921,79721,949
Mean21,46621,90821,31321,34221,461
Worst20,84121,85520,95720,95820,886
cb4-4100 × 1022,772Best22,31922,71722,32522,41822,376
Mean21,99222,52821,96122,03722,029
Worst21,46522,01621,48821,22821,533
cb4-5100 × 1022,751Best22,44022,75122,16822,21522,254
Mean22,13222,60321,84021,82221,903
Worst21,73822,27221,27121,36221,339
cb5-1250 × 1059,187Best57,75758,90356,92857,53057,036
Mean56,70858,47755,75955,85455,960
Worst55,51058,18254,21753,57053,381
cb5-2250 × 1058,781Best57,36358,34656,33756,56856,490
Mean56,79358,09855,45555,44355,708
Worst56,12657,79253,73953,27452,907
cb5-3250 × 1058,097Best56,69057,67455,57356,42655,982
Mean56,02457,41754,63854,79354,727
Worst55,28157,04453,51652,87152,714
cb5-4250 × 1061,000Best59,93060,50558,59559,03059,077
Mean58,93460,28257,76658,05757,721
Worst57,76559,87056,70156,25453,774
cb5-5250 × 1058,092Best56,86357,46856,18656,21756,204
Mean56,06657,22054,85054,94154,872
Worst55,18256,86953,61251,85050,832
cb6-1500 × 10117,821Best113,362116,015108,487110,996111,669
Mean112,541115,379105,760107,698108,367
Worst111,397114,509102,725104,239103,802
cb6-2500 × 10119,249Best115,022117,778109,569114,262113,001
Mean114,250117,102106,775108,648109,197
Worst112,596116,418103,478100,740100,764
cb6-3500 × 10119,215Best115,419117,345109,705113,987112,419
Mean114,372116,842106,853108,576109,004
Worst113,495116,115104,565102,439103,703
cb6-4500 × 10118,829Best115,038117,281108,628112,476112,198
Mean113,444116,446105,679107,692107,796
Worst112,405115,872102,679101,86099,470
cb6-5500 × 10116,530Best112,971114,909106,972109,567109,287
Mean111,707114,183104,509106,217106,212
Worst110,156113,533102,665100,836100,509
#Best 015000
#Mean 015000
#Worst 014000
AE of Best 2.30%0.76%4.58%3.25%3.52%
AE of Mean 3.57%1.33%6.58%5.92%5.78%
AE of Worst 5.20%2.12%8.77%9.64%10.11%
Rank of AE of Best21534
p-value (Mean)0.000 0.0000.0000.000
Table 6. The results of HLMS with 6 comparative algorithms for Test set IV.
Table 6. The results of HLMS with 6 comparative algorithms for Test set IV.
Prob.n × mOpt. MSHLMSBIWOABMMVOBSCABHHABSSA
GK01100 × 153766Best3732375237433698372537463744
Worst3714372237313666369037283738
Mean3721374237363678370537343742
GK02100 × 253958Best3920394839493885391339293939
Worst3900392839243859388339153924
Mean3908393839313871389739193934
GK03150 × 255656Best5585561056135561556355805606
Worst5545557555845507553355545594
Mean5564559655965519554355685598
GK04150 × 505767Best5702573357125651567856955712
Worst5677565856905628565256725696
Mean5688571057015638566456835704
GK05200 × 257561Best7479750274997365741174637495
Worst7422741374767344737574267477
Mean7452747474857353739174437488
GK06200 × 507680Best7617761176077522755175787617
Worst7561756875847492752075627598
Mean7591759275977505753275697611
GK07500 × 2519,220Best19,06619,15119,11018,73818,78319,00519,100
Worst19,00518,89019,09318,63518,68918,96119,048
Mean19,03319,06719,10218,66818,73418,98319,087
GK08500 × 5018,806Best18,61218,64218,64118,38518,46218,60118,646
Worst18,55718,49818,60718,33518,39518,59718,637
Mean18,58218,59418,62718,36118,42818,59818,640
GK091500 × 2558,087Best57,75357,88657,86856,74656,93257,54757,346
Worst57,63656,72557,83056,51956,62456,69956,615
Mean57,67657,54757,84356,61956,71957,71956,959
#Best 1520002
#Worst 0120006
#Mean 0320005
AE of Best 0.11%0.07%0.08%0.24%0.19%0.12%0.09%
AE of Mean 0.16%0.18%0.11%0.29%0.27%0.17%0.13%
AE of Worst 0.13%0.10%0.10%0.27%0.22%0.13%0.11%
Rank on AE of Best 4127653
p-value (Mean) 0.327 0.1720.0010.0010.0690.327
Table 7. The results of HLMS using different parameters of two distributions for Test set I.
Table 7. The results of HLMS using different parameters of two distributions for Test set I.
Prob.   μ = 0.5 , δ = 0.3 x = 0, y = 1x = 0, y = 0.5x = 0, y = 2x = −2, y = 1
Sento1Mean77297740774977537749
SR0.430.670.630.670.60
Sento2Mean87018713870887128713
SR0.130.330.100.070.27
HP1Mean33913376337133853367
SR0.430.200.030.070.10
HP2Mean31153125312431353119
SR0.100.030.070.000.00
PB1Mean30453058306030573055
SR0.130.130.330.170.20
PB2Mean31453145314231483145
SR0.070.130.100.100.10
PB4Mean92,09392,66492,80893,22992,457
SR0.000.100.070.030.03
PB5Mean20982118211221152106
SR0.170.500.470.570.37
PB6Mean753757757756758
SR0.430.530.530.530.53
PB7Mean10241026102510251026
SR0.100.130.130.070.13
Weing1Mean139,551139,615139,397140,199138,803
SR0.330.070.230.200.13
Weing2Mean130,307130,370126,292129,613129,989
SR0.270.400.230.200.60
Weing3Mean92,36294,54693,37393,03394,489
SR0.270.500.500.300.60
Weing4Mean118,384117,864117,199117,489117,544
SR0.230.200.130.030.13
Weing5Mean96,46595,42197,20995,37294,898
SR0.430.330.670.400.37
Weing6Mean129,533128,954129,753129,983129,732
SR0.370.170.400.370.47
Weing7Mean1,048,3781,038,3661,043,5951,035,0591,035,752
SR0.000.000.000.000.00
Weing8Mean622,201621,720622,515621,778622,127
SR0.000.000.030.030.00
#Mean 25363
#SR 57656
MSR 0.220.250.260.210.26
p-value (Mean) 1.0000.5560.7270.635
Table 8. Comparisons of MS, HLMS-B, HLMS-H, and HLMS on Test set I.
Table 8. Comparisons of MS, HLMS-B, HLMS-H, and HLMS on Test set I.
Prob. MSHLMS-BHLMS-HHLMS
Sento1Mean7677774977637762
SR0.170.630.870.90
Sento2Mean8695870887178719
SR0.000.100.530.47
HP1Mean3370337133923399
SR0.130.030.370.40
HP2Mean3098312431353158
SR0.000.070.300.40
PB1Mean3036306030753074
SR0.030.330.500.43
PB2Mean3139314231723178
SR0.000.100.600.70
PB4Mean92,31292,80893,11193,063
SR0.000.070.170.30
PB5Mean2091211221172125
SR0.170.470.570.70
PB6Mean751757767768
SR0.430.530.830.80
PB7Mean1023102510281030
SR0.030.130.200.50
Weing1Mean141,207139,397141,227141,260
SR0.900.230.830.93
Weing2Mean130,760126,292130,877130,877
SR0.300.230.970.97
Weing3Mean90,86693,37395,35594,801
SR0.130.500.830.80
Weing4Mean116,487117,199118,883118,956
SR0.730.130.500.37
Weing5Mean95,80297,20998,38498,796
SR0.230.670.871.00
Weing6Mean129,176129,753130,429130,610
SR0.230.400.630.97
Weing7Mean1,069,1211,043,5951,073,4671,073,783
SR0.000.000.000.00
Weing8Mean620,483622,515622,905622,775
SR0.330.030.000.03
#Mean 00613
#SR 12713
MSR 0.210.260.530.54
p-value 0.00030.00030.155
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, Y.; Wang, H.; Cai, Z.; Li, M.; Li, X. Hybrid Learning Moth Search Algorithm for Solving Multidimensional Knapsack Problems. Mathematics 2023, 11, 1811. https://doi.org/10.3390/math11081811

AMA Style

Feng Y, Wang H, Cai Z, Li M, Li X. Hybrid Learning Moth Search Algorithm for Solving Multidimensional Knapsack Problems. Mathematics. 2023; 11(8):1811. https://doi.org/10.3390/math11081811

Chicago/Turabian Style

Feng, Yanhong, Hongmei Wang, Zhaoquan Cai, Mingliang Li, and Xi Li. 2023. "Hybrid Learning Moth Search Algorithm for Solving Multidimensional Knapsack Problems" Mathematics 11, no. 8: 1811. https://doi.org/10.3390/math11081811

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop