Next Article in Journal
Assessment of Heavy Metal and Oil-Contaminated Silty Sand Treatment by Electrical Resistance Heating Method
Next Article in Special Issue
Three-Dimensional Formation Control for Robot Swarms
Previous Article in Journal
Thermal Performance Analysis of Gradient Porosity Aluminium Foam Heat Sink for Air-Cooling Battery Thermal Management System
Previous Article in Special Issue
Interoperability between Real and Virtual Environments Connected by a GAN for the Path-Planning Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Trajectory Generation and Optimization Using the Mutual Learning and Adaptive Ant Colony Algorithm in Uneven Environments

1
College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China
2
School of Computer Science and Engineering, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(9), 4629; https://doi.org/10.3390/app12094629
Submission received: 19 March 2022 / Revised: 29 April 2022 / Accepted: 30 April 2022 / Published: 4 May 2022
(This article belongs to the Special Issue Advances in Robot Path Planning)

Abstract

:
Aiming at the trajectory generation and optimization of mobile robots in complex and uneven environments, a hybrid scheme using mutual learning and adaptive ant colony optimization (MuL-ACO) is proposed in this paper. In order to describe the uneven environment with various obstacles, a 2D-H map is introduced in this paper. Then an adaptive ant colony algorithm based on simulated annealing (SA) is proposed to generate initial trajectories of mobile robots, where based on a de-temperature function of the simulated annealing algorithm, the pheromone volatilization factor is adaptively adjusted to accelerate the convergence of the algorithm. Moreover, the length factor, height factor, and smooth factor are considered in the comprehensive heuristic function of ACO to adapt to uneven environments. Finally, a mutual learning algorithm is designed to further smooth and shorten initial trajectories, in which different trajectory node sequences learn from each other to acquire the shortest trajectory sequence to optimize the trajectory. In order to verify the effectiveness of the proposed scheme, MuL-ACO is compared with several well-known and novel algorithms in terms of running time, trajectory length, height, and smoothness. The experimental results show that MuL-ACO can generate a collision-free trajectory with a high comprehensive quality in uneven environments.

Graphical Abstract

1. Introduction

Mobile robots are increasingly used in outdoor applications, such as search and rescue missions, planetary ground surveys, and national defense security [1] where robots usually have to face complex environments with an uneven terrain [2,3,4]. In such an environment, the planned trajectory usually includes path segments with rapidly changing height and multiple sharp turns. When the mobile robot passes through this section of the path, it will consume more energy [5], which should be avoided. Therefore, this paper focuses on trajectory generation and optimization to quickly find a short and smooth trajectory in uneven environments.
In an uneven environment, it is obvious that ordinary 2D grid maps cannot reasonably model the environment, so it is essential to build a map model that can better describe the uneven environment. Generally, 3D modeling and 2.5D elevation maps are used in this circumstance. In [6,7,8], 3D radar for 3D modeling was used which simulates the complex uneven outdoor environment including stairs, slopes, etc. This method can accurately and comprehensively grasp the specific information of the surrounding environment [9], but it occupies a large amount of computer running memory and has poor timeliness, especially when the size of the environment increases. Therefore, scholars have tried to find simpler and more applicable modeling methods. In [10,11,12], the authors tried to reduce the cost of 3D radar modeling, using RGB-D (Red Green Blue-Depth) sensors to store the height information and coordinates of the environment in the same grid to obtain a discrete 3D grid map, namely a 2.5D elevation map. This paper proposes a 2D-H grid map based on a 2.5D elevation map, which maps the height information of the environment to a 2D plane. Compared with the 2.5D elevation map, the 2D-H method further reduces the number of calculations and saves memory space, which not only meets the needs of timeliness but is also practical enough.
The goal of trajectory planning is to find a trajectory for a robot from a starting point to an ending point without colliding with obstacles and meeting other constraints, such as the trajectory length and smoothness. In recent years, trajectory planning has attracted much attention in the field of robotics. Traditional methods, such as simulated annealing (SA) [13] and the artificial potential field method (APF) [14], are widely used because of their ease of understanding and implementation. However, [15] these methods may easily fall into a local optimal solution and cannot reach the end point when a complex environment is encountered.
When dealing with trajectory planning in a complex dynamic environment, intelligent bionic algorithms often play an important role [16], including genetic algorithms (GA) [17], neural networks [18], particle swarm optimization (PSO) [19], and the ant colony algorithm (ACO) [20]. The GA algorithm has strong uncertainty in the planned trajectory due to its large number of parameters and the randomness of the initial population [21]. In the PSO algorithm, [22] the initial particles are randomly selected, which leads to a variety of different planning results. [23]. Neural network algorithms spend a lot of time on pre-training and adjusting parameters. Not only is the number of calculations significant, but also the interpretability is not ideal. In an uneven environment with obstacles, the uncertainty of the environment, such as the distribution of obstacles and the complexity of the environment, brings difficulties to the trajectory planning of mobile robots. Therefore, the required path planning algorithm needs a strong search ability and good stability.
ACO is a probabilistic technique to solve computational problems which is robust and easy to combine with other methods [24,25]. In recent years, the ant colony algorithm has been widely used in transportation, logistics distribution, network analysis, and other fields [26]. Nevertheless, the traditional ant colony algorithm has the defects of a low search efficiency and a slow convergence speed, and it is also easy to fall into local extremes.
Researchers have put forward various improved methods to optimize the search ability of the ant colony algorithm. Li et al. [27] proposed an improved ant colony algorithm with multiple heuristics (MH-ACO), which is better reflected in the global search ability and convergence. However, the parameter setting is complex, which brings randomness to the experiment. Akka et al. [28] used a stimulating probability to help the ants choose the next grid and employed new heuristic information based on the principle of unlimited step length to expand the field of view and improve the visibility accuracy. The improved algorithm speeds up the convergence speed and expands the search space, but the safety of the trajectory cannot be guaranteed. In addition, Ning et al. [29] designed a new pheromone smoothing mechanism to enhance the search performance of the algorithm. When the search process of the ant colony algorithm is close to stagnation, the pheromone matrix is reinitialized to increase the diversity of the connections at the expense of a large time cost.
Furthermore, the improved ant colony algorithm is combined with some two-stage trajectory planning methods. Chen et al. [30] proposed a fast two-stage ant colony algorithm based on the odor diffusion principle, including two stages of preprocessing and trajectory planning, which accelerated the convergence speed but did not consider the trajectory optimization. Yang et al. [31] proposed a multi-factor improved ant colony algorithm (MF-ACO) to solve the problem related to the fact that the trajectory planning algorithm of mobile robots cannot cope with the complex actual environment. The maximum and minimum ant strategy was adopted to avoid local optima. Then the dynamic tangent point adjustment method was used to smooth the path to further improve the quality of the trajectory, but the smoothness needed to be further improved.
Although the above-mentioned improved algorithm attempts to improve the search performance of ACO and speed up the convergence speed, it does not consider the height information in the environment, which is different from ordinary raster maps, and the convergence performance of ACO can be further improved. For trajectory planning and optimization in complex and uneven environments, this paper proposes a trajectory generation and optimization method based on mutual learning and adaptive ant colony optimization (MuL-ACO), which can make robots safely and quickly plan a short and smooth trajectory in uneven environments. In this method, the global trajectory planning of mobile robots is divided into two consecutive parts: the initial trajectory generation and trajectory optimization. Firstly, the improved adaptive ant colony algorithm further accelerates the convergence of ACO to quickly generate the initial trajectory. The initial planned trajectory may contain redundant points and inflection points, which results in high memory consumption and poor trajectory quality. Then, a trajectory optimization algorithm based on mutual learning is proposed to further optimize the length and smoothness of the initial trajectory.
The main contributions of this paper are as follows.
  • The 2D-H raster map is proposed to simulate the uneven outdoor environment. The height information in the three-dimensional environment is stored in the 2D plane;
  • A hybrid scheme using mutual learning and adaptive ant colony optimization is proposed in this paper. The global robot trajectory planning problem is divided into two consecutive parts: the trajectory generation part and the trajectory optimization part;
  • An improved adaptive ant colony algorithm is proposed to generate the initial trajectory. Considering the height information of the map, a comprehensive heuristic function including length, height, and smoothness is designed. Then, a new pheromone adaptive update strategy is proposed through an improved simulated annealing function to speed up the convergence of the algorithm.
  • A new trajectory optimization algorithm based on mutual learning is proposed to optimize the generated initial trajectory. Firstly, feature ablation experiments are carried out for each turning point to obtain the safety feature sequence of each turning point. Then, each point learns from other points to gradually eliminate the points that do not affect the trajectory safety to optimize the trajectory length and smoothness. Finally, the shortest sequence of key points affecting trajectory safety is obtained. Therefore, the algorithm optimizes the final trajectory in terms of smoothness, length, and stability.
This research is structured as follows: Section 2 describes the environment and problems. Section 3 illustrates the improved adaptive ant colony algorithm. Section 4 describes the framework and the process of the proposed trajectory optimization algorithm based on mutual learning in detail. Section 5 presents the steps and flowcharts of a hybrid scheme using mutual learning and adaptive ant colony optimization. Section 6 discusses the results of the simulation experiment. Finally, Section 7 concludes the paper.

2. Environment Description and Problem Formulation

2.1. Environment Description

In an uneven environment, the working environment of a mobile robot is a 2D-H grid map. The grid of the map is divided into grids occupied by obstacles, grids with height information, and free grids.
In the 2D-H map, the obstacle grid is considered impassable, the free grid is considered passable, and the height grid affects the quality of trajectories planned by robots. The height grid is modeled by a normalized method, as shown in Equation (1). H ( i ) is the color intensity value of the i-th height grid cell, and the color becomes darker as the value of H ( i ) increases, H ( i ) [ 0 , 1 ] . h is a height function when h ( i ) > 0 and H represents the intensity value of green, otherwise H represents the intensity value of blue.
H ( i ) = { max ( h ) h ( i ) max ( h ) , h ( i ) > 0 min ( h ) h ( i ) min ( h ) , h ( i ) < 0
Figure 1 shows the modelling process of the 2D-H map. The uneven environment here is simulated with the peaks function, a probability density function of a binary Gaussian distribution. Researchers can design other functions to simulate an uneven environment according to different requirements. Figure 1a shows the three-dimensional model of the mountain function and the contour mapping on the 2D plane. Figure 1b shows the 2D-H map generated by the peak function and obstacles, where black grid cells represent obstacles, blue cells represent concave areas, and green cells represent raised areas. Here the height information in the map and the location of static obstacles are known.

2.2. Problem Formulation

Based on the 2D-H map, the problem of trajectory generation and optimization is described as follows. Given a starting point and an ending point, a robot is expected to plan an initial trajectory between them. In the trajectory generation stage, the optimization is to minimize the length, height difference, and number of turns of the trajectory. Then, in the trajectory optimization stage, redundant nodes of the initial trajectory are further reduced to obtain a shorter and smoother trajectory.
The robot can only move one grid distance in a time step, the side length of one grid is set to 1 m, and the height threshold is set to (−1 m, 1 m) to limit the movement of the robot. The mobile robot is regarded as a mass point and moves at a fixed speed. As shown in Figure 2, there are eight moving directions of the robot, which point from the center of the current grid to the center of the adjacent eight grids. Figure 2b shows a trajectory of the robot moving from the starting point to the target point. The planned trajectory is limited by the height threshold, so the darkest blue and green grids are bypassed.

3. Improved Adaptive Ant Colony Algorithm

3.1. Basic Ant Colony Optimization

Ant colony optimization is a metaheuristic algorithm inspired by the behavior of ants in nature. It imitates the foraging behavior of ant colonies to find the optimal trajectory in an unknown environment. Two factors determine the next step of ants, which are heuristic information and pheromones. Heuristic information is obtained from the surrounding environment, and pheromones are the directional information emitted by the group. In the process of food searching, the ants release pheromones on the trajectory they walked by, which will attract other ants. As the number of ants on the same trajectory increases, pheromones are gradually accumulated, and more ants are attracted to the trajectory. According to the surrounding environment and experience, the transition probability of ants can be calculated by Equation (2).
p i , j k ( t ) = { [ τ i , j ( t ) ] a [ η i , j k ( t ) ] β s   A i [ τ i , j ( t ) ] a [ η i , j k ( t ) ] β , j   A   i 0 , j   A   i
where p i , j k ( t ) is the transition probability of the k-th ant from point i to point j at the t-th iteration, τ i , j ( t ) is the pheromone concentration from point i to point j, η i , j k ( t ) is the heuristic function of the k-th ant from point i to point j, A i is a set of feasible points adjacent to j, and α and β are weights that control the relative importance of τ i , j and η i , j . If the proportion of α is larger, then ants tend to choose the trajectory that most of the previous ants have walked. If the proportion of β is larger, ants tend to choose the trajectory to traverse based on the heuristic information in the environment. Heuristic function and pheromone concentration directly affect the behavior of ants searching for trajectories. The heuristic function η i , j k ( t ) can be calculated by Equation (3), where d j g is the Euclidean distance between the next node j and the end point g.
η k i , j ( t ) = 1 d j g
When all the ants complete the trajectory search, the pheromones will accumulate on the passing trajectory. The pheromone τ i , j ( t ) is updated once per iteration, as shown in Equation (4).
τ i , j ( t + 1 ) = ( 1 ρ ) τ i , j ( t ) + k = 1 m Δ τ i , j k ( t )
Δ τ i , j k ( t ) = { Q / L k , if   ant   k   visits   i   and   j 0 , otherwise
Here Δ τ i , j k ( t ) is the pheromone increment left by ant k after passing the trajectory between i and j, at the t-th iteration. ρ is the pheromone volatilization factor, which adjusts the accumulation speed of the pheromone. Moreover, m is the total number of ants. Generally, Δ τ i , j ( 0 ) = C , C is the initial pheromone constant. Q is a constant, and L k is the total length of the trajectory passed by ant k in this iteration, which is inversely proportional to Δ τ i , j k ( t ) .

3.2. Improved Heuristic Function

To reduce the blindness of the ants in the early stage and to find a short and smooth trajectory in an uneven environment, the heuristic function is redesigned. The smoothness and height factors are added to the heuristic function, as shown in Equation (6). The comprehensive heuristic function includes length d, a smoothing function s, and a height function h.
η i , j ( t ) = λ d ( i , j , g ) + γ h ( i , j ) + ψ s k i , j ( t )
In the heuristic function, the distance function d ( i , j , g ) , height function h ( i , j ) , and smoothing function s i , j k ( t ) are defined as follows. λ , γ , ψ are weight parameters and they depend on the application.
d ( i , j , g ) = d max ( j , g ) d j , g d max ( j , g ) d min ( j , g ) + 0.01 × ω
where d ( i , j , g ) is the corrected distance from a certain adjacent grid cell j of the current grid cell i to the target grid cell. d ( i , j , g ) is designed to enlarge the distance gap between adjacent grid cells. d max and d min are the maximum and minimum distances between adjacent grid cells and the end point. w is a correction parameter, and 0.01 avoids the situation where the denominator is 0.
h ( i , j ) = { h max | h ( i ) h ( j ) | h max h min + 0.01 × ω , | h ( i ) h ( j ) | h c   , | h ( i ) h ( j ) | > h c
where h ( i , j ) is the corrected height from the current grid cell i to a certain neighboring grid cell j. h ( i , j ) is designed to guide ants to visit flatter grid cells. h c is the height constraint value of the robot, which should meet the limit of the height threshold. h max and h min are the maximum and minimum height differences between adjacent grid cells and the current grid cell. w is a correction parameter, and 0.01 avoids the situation where the denominator is 0.
s i , j k ( t ) = { u , d i r a , i k ( t ) = d i r i , j k ( t ) 0.1 u , otherwise
where s i , j k ( t ) is the smoothness of the trajectory passed by the k-th ant in the t-th iteration. Generally, ants can move in eight directions adjacent to the current grid. Suppose an ant reaches the current point i from a, and then goes to the next point j. If the moving direction of the previous step d i r a , i k is consistent with the moving direction of d i r i , j k , the reward u will be given, and the value of u is set to 5 in this paper.

3.3. SA-Based Adaptive Adjustment Strategy of Pheromones

In the traditional ant colony algorithm, the pheromone is often a fixed value and will not automatically adjust with the iteration situation. Ants search blindly in the early stage due to the low concentration of pheromones. In the later stage, pheromones will accumulate in a large amount, which will reduce the diversity of trajectory selection. Therefore, the fixed increase in the pheromones cannot meet the requirements of search efficiency and will slow down the convergence speed.
Based on the temperature update function of the simulated annealing algorithm, this paper proposes a new annealing strategy to dynamically update the pheromone volatilization factor. The temperature T is an important control parameter in the simulated annealing algorithm, which determines the annealing direction and the running speed of the algorithm. As the iteration makes the de-temperature function have linear characteristics, the running speed of the de-temperature function is further improved. Moreover, a dynamic pheromone volatilization factor ρ is designed to realize the adaptive update of pheromone. ρ is set by Equations (10) and (11).
ρ = l exp μ / T i
T ( i ) = T end + T start T end I max I i I max , I i I max 2 T end + T start T end I max I max I i I max , I i > I max 2
where μ is a constant and T i is the current temperature, which is calculated by the de-temperature function T ( i ) . I max is the number of maximum iterations. T start is the initial temperature, and T end is the final temperature. As the number of iterations increases, the de-temperature function linearly decreases and then rises to further increase the running speed of the de-temperature function. To achieve better experimental results, the value of parameter μ is discussed in this paper. Other parameters are set to: T start = 100 , T end = 0.1 , I max = 50 .
In the previous work, many experiments were performed to determine the value of μ, and finally three possible and suitable values were selected and are discussed in this paper. The values of μ were set to 33, 50, and 90, respectively, to calculate ρ according to Equation (10). The adaptive change curve of ρ is shown in Figure 3. In the early stage, the volatilization factor ρ is small, which is conducive to accumulating pheromones and improving the directionality of the ants’ search. In the mid-term, ρ becomes larger to speed up the iteration speed of the algorithm and to avoid falling into the local optimum. In the later stage, ρ becomes small to accelerate convergence.
To further test the influence of the value of μ on the convergence of the improved adaptive ant colony algorithm (IAACA) proposed in this paper, three benchmark functions were applied for verification. Other swarm intelligence algorithms including ACO, PSO, Grey Wolf Optimizer (GWO) [32] were chosen to compare with IAACA in terms of convergence performance. Table 1 presents the details about the different types of benchmark functions. The three benchmark functions had a minimum value of 0, the particle dimension was set to 10, and the value range of x and y was (−5.12, 5.12). Sphere is a unimodal function, while Rastrigin and Ackley are multimodal functions with many local values and difficult to solve. These three benchmark functions were suitable for testing the convergence performance of the algorithm.
The convergence curves of PSO, GWO, ACO, and IAACA (μ = 33, μ = 50, μ = 90) corresponding to the three benchmark functions are shown in Figure 4. It can be seen from Figure 4 that the convergence performance of IAACA was the best when μ = 33. As shown in Figure 4a, the convergence performance of IAACA (μ = 33) was better than the other three algorithms. Although GWO converged rapidly downward, it still did not converge in the end. In Figure 4b,c, IAACA (μ = 33) and GWO converged at around 50 iterations, and the convergence performance was better than the other two algorithms. Figure 4b,c showed that IAACA (μ = 33) and GWO had a better convergence performance, but IAACA (μ = 33) was more stable from Figure 4a.

4. Mutual Learning-Based Trajectory Optimization Algorithm

To reduce the redundant nodes of the initial trajectory generated by the adaptive ant colony algorithm to further optimize the length and smoothness of the trajectory, a trajectory optimization algorithm based on mutual learning is proposed. Firstly, feature ablation experiments were carried out for each turning point to obtain the safety feature sequence of each turning point. Then, each point learns from other points to gradually eliminate the points that do not affect the trajectory safety to optimize the trajectory length and smoothness. Finally, the shortest sequence of key points affecting trajectory safety is obtained. The proposed algorithm achieved a smooth trajectory and minimized the length of the trajectory. At the same time, the wear of the robot’s steering to follow the planned trajectory was reduced.
For instance, as shown in Figure 5, the feasible initial trajectory L o l d : S N 1 N 8 E is usually not the best trajectory. After the trajectory optimization algorithm based on mutual learning is optimized once, the trajectory L 1 can reach the end point without passing through the N 1 point. When the optimization is completed, the final trajectory L b e s t can even reach the end point directly through the N 8 point. Consequently, L o l d is optimized to L b e s t : S N 8 E , which optimizes the length and smoothness of the initial trajectory. The algorithm is described as follows.
The initial trajectory L o l d , including the starting point S and the ending point E, can be represented by a set of all turning points N = { S , N 1 , N 2 , , N i , , N n , E } , i [ 1 , n ] , and the coordinates of the point set are represented by Equation (12).
R = { ( x 0 , y 0 ) , ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x i , y i ) , , ( x n , y n ) , ( x n + 1 , y n + 1 ) }
To learn the collision characteristic of each turning point, it is assumed that there are n initial individuals P i represented by the characteristic matrix (13). Then each individual subjected to a characteristic ablation experiment, and the characteristic zero point is set by Equation (14).
P i = [ N 0 , N 1 , N 2 , , N i , , N n , N n + 1 ]  
N i = { 0 , N i   is   ablated 1 , otherwise
A reward and punishment step are added to determine whether the i-th individual P i reaches the target directly from the starting point without collision. The cost function is L i ( P i ) calculated by Equation (15), L o l d is the length of the initial trajectory, and L i is the trajectory length of the current individual.
L i ( P i ) = { L i ,   i f   P i   n o   c o l l i s i o n   L o l d ,   o t h e r w i s e  
L i = k = 0 n + 1 d N k , N k + 1 = d N 0 , N 1 + d N 1 , N 2 + + d N k 1 , N k + d N k , N k + 1 + + d N n , N n + 1 , i f ( N r = 0 ) = d N 0 , N 1 + d N 1 , N 2 + + d N r 1 , 0 + d 0 , N r + 1 + + d N n , N n + 1 = d N 0 , N 1 + d N 1 , N 2 + + d N r 1 , N r + 1 + + d N n , N n + 1
d N k , N k + 1 = ( x k x k + 1 ) 2 + ( y k y k + 1 ) 2
The current individual P i learns from each other in turn with other individuals P j , and learns the collision characteristics of each other’s nodes to obtain the best individual. Suppose P 1 is the current individual, the mutual learning process is shown in Figure 6.
In the first stage of initialization, each turning point is subjected to an ablation experiment to obtain the feature and feature sequence of the point, which can be obtained by Equation (14). In the second stage of mutual learning, individuals start from P 1 and learn from each other in turn. The learning method is to compare the value of the cost function, and individuals with high values learn from individuals with low values. In the third stage of the individual update after mutual learning, if the current individual penalty value L i does not increase, the old individual P i is updated to the new individual P · i according to the cost function, otherwise, it is not updated. The individual is updated by Equation (18).
P ^ i = { P ˙ i ,   i f   L ^ i ( P ˙ i ) < L i ( P i ) P i ,   o t h e r w i s e  
The ultimate goal of the mutual learning trajectory optimization algorithm is to generate an optimized path with fewer turns D best and a shorter length L best .
L best   = min { L ^ 1 , L ^ 2 , , L ^ i , , L ^ n }
D best   =   length   { P best   0 }
Algorithm 1. The pseudo code for MLTO.
Algorithm 1. Mutual learning-based Trajectory Optimization Algorithm
1: input turning point set N
2: initialize point set N as feature individual P i by (13)(14)
3: calculate the reward and punishment function L i by (15)(16)(17)
4: for  i = 1 to n do
5:  for  j = 1 to n do
6:    if  L j is not more than L i then
7:      P i learns feature zero through P j , and calculate L ^ i
8:      P i will be updated to P ^ i by (18)
9:    else
10:      P j learns feature zero through P i , and calculate L ^ j
11:      P j will be updated to P ^ j by (18)
12:    end if
13:  end for
14: end for
15: if  L b e s t is equal to the initial trajectory L then
16:  the optimal trajectory is initial trajectory L , number of turns is D
17: else 
18: calculate the length of the optimal trajectory and the number of turns by (19)(20)
19: end if
20: output the number of turns: D b e s t
21: output the shortest trajectory: L b e s t

5. A Hybrid Scheme Using MuL-ACO for Trajectory Generation and Optimization

For the robot to effectively find a short and smooth trajectory while avoiding crossing steep areas, a hybrid scheme using mutual learning and adaptive ant colony, namely, MuL-ACO, is presented in this work. Figure 7 shows the flow chart of the scheme.
Step 1:
Establish a 2D-H grid map based on the uneven environment and initialize the parameters. Set the starting point S, the end point E, the number of ants k, the maximum number of iterations I max , pheromone heuristic factor α, expected heuristic factor β, pheromone intensity coefficient Q, and the initial pheromone τ ( 0 ) .
Step 2:
Trajectory selection. Calculate the heuristic function by Equation (6). The ant is placed at the starting point and the probability of transferring to the next node is calculated by Equation (2). All trajectory nodes from the starting point to the current point are stored in the Tabu list.
Step 3:
Determine whether all the ants have completed the trajectory search in this generation. If it is, go to step 4, otherwise, return to step 2.
Step 4:
Record the nodes of the trajectory walked by all ants and find the optimal trajectory in this iteration.
Step 5:
Adjust the pheromone volatilization factor ρ adaptively, according to Equations (10) and (11). Update the pheromone by Equation (4). Re-zero the taboo table.
Step 6:
Determine whether the maximum number of iterations I max is reached. If it is, go to step 7, otherwise, return to step 2.
Step 7:
Initial trajectory generation. Obtain the optimal trajectory node and total height.
Step 8:
Trajectory optimization. The trajectory optimization algorithm based on mutual learning, such as Algorithm 1, is used to optimize the initial trajectory. Calculate the optimal trajectory including length, height, and number of turns.

6. Experiment

In this section, four sets of simulation experiments were conducted to evaluate the performance of the MuL-ACO scheme. In the first set of experiments, the algorithm proposed in this paper was compared with other intelligent algorithms, which are the well-known GA and the novel sparrow search algorithm (SSA) [33]. In the second set of experiments, the algorithm proposed in this paper was compared with the traditional ACO and other improved ACO, which are MH-ACO and MF-ACO. Two groups of experiments were carried out on maps with different sizes and different numbers of obstacles. The third set of experiments was set with different starting points and ending points on a map, and compared with the proposed algorithm with ACO, MH-ACO, and MF-ACO. To further verify the effectiveness of the improved algorithm, the fourth set of experiments was simulated in a dynamic environment. Furthermore, the height threshold of uneven environments was set to (−1 m, 1 m) to constrain the robot’s trajectory search. The initial trajectory is given on each map to show the process of trajectory generation and optimization.
To build the environment map of the mobile robot, 2D-H grid maps with different sizes, obstacles, and terrains were modeled by the MATLAB simulation platform, where the blue areas are low terrain areas, and the green areas are high terrain areas. Obstacles were randomly placed on the map and start and end points were set. All experiments were performed on the same PC to obtain an unbiased comparison of CPU time. Parameters of MuL-ACO were set as the following: k = 50 ,   I max = 50 , α = 3 ,   β = 6 , C = 10 ,   Q = 100 ,   u = 5 , w = 8 ,   h c = 1 , τ ( 0 ) = 20   , μ = 33 , λ = γ = ψ = 1 .

6.1. Simulation Experiment A

In this experiment, six maps of 20 × 20 m were selected for simulation, which differed in the number of obstacles and the shapes of obstacles. Figure 8 shows the optimal trajectories planned by GA, SSA, and MuL-ACO. To obtain more specific performance indicators, including length, number of iterations, smoothness, height difference, and running time, the experiment was performed 30 times to obtain the average value. Table 2 summarizes the qualitative comparison of the performance of MuL-ACO, GA, and SSA.
In Figure 8a–f, the trajectory planning based on MuL-ACO tended to bypass steep or concave areas and generated a comprehensive optimal target trajectory with good safety, fewer turns, and a shorter length. The main reason is that MuL-ACO introduces the height factor in environments into the heuristic function, and the trajectory optimization algorithm based on mutual learning further optimizes the trajectories. In contrast, the GA and SSA algorithms do not actively bypass steep or concave areas when generating trajectories, so mobile robots will incur certain losses when following trajectories and the quality of trajectories is average. In the case of a small difference in trajectory length, as shown in Figure 8a,b,e, MuL-ACO can find a flatter and safer trajectory through the trajectory optimization algorithm based on mutual learning. In Figure 8c,d,f, the trajectories generated by MuL-ACO were shorter, smoother, and safer than GA and SSA.
As shown in Table 2, compared with GA and SSA, the hybrid scheme of MuL-ACO proposed in this paper had a large improvement in the convergence performance, which was about 70% to 85%. The reason is that the SA-based pheromone adaptive adjustment strategy accelerated the convergence of the algorithm. In terms of the height difference of the trajectory, the trajectory planned by MuL-ACO tended to avoid passing through steep uneven areas, so the total height difference of the trajectory was lower than GA and SSA. Compared with the other two algorithms, the final trajectory generated by the mutual learning-based trajectory optimization algorithm was better in terms of length and smoothness. In addition, it can be seen from Table 2 that MuL-ACO had a shorter running time when planning the trajectory.

6.2. Simulation Experiment B

In this experiment, six maps of 30 × 30 m were selected for simulation, which differed in the number of obstacles and the shapes of obstacles. Start and end positions were given randomly, as shown in Table 3. MuL-ACO was compared with ACO, MH-ACO, and MF-ACO in larger and more complex uneven environments.
Figure 9 shows the best trajectory planned by each method. It is obvious that the trajectory planned by the method proposed in this paper had fewer turns, good smoothness, and short length, while bypassing steep or concave terrain. In simple scenes, as shown in Figure 9c,d,e, the trajectory planned by MuL-ACO was smoother and bypassed steep or concave areas as much as possible. In more complex scenes (Map 1 and Map 6), the difference in the quality of the trajectories was more obvious. As shown in Figure 9a, the trajectories planned by ACO and MH-ACO extended along the edge of the obstacle resulting in multiple turns. Although the MF-ACO algorithm avoided local traps when planning trajectories, its trajectory height and smoothness were worse than MuL-ACO. Especially in Vortex map 6, as shown in Figure 9f, the traditional ACO failed to generate a trajectory from the starting point to the end point, and MH-ACO often fell into a local trap and could not escape. Although MF-ACO could work normally, the quality of the trajectory was not ideal. MuL-ACO had a good performance due to the improved heuristic function and adaptive pheromone adjustment strategy. Moreover, the trajectory optimization algorithm based on mutual learning is suitable for different maps. Therefore, the scheme proposed in this paper is not affected by changeable environments and can plan the trajectory with a good comprehensive quality.
As shown in Table 4, 30 experiments were performed on six different maps to obtain the specific average performance metric of the algorithms. Figure 10 shows the performance comparison of MuL-ACO and the other three algorithms on 6 maps, which more clearly highlights the advantages of the proposed scheme in terms of length, number of iterations, number of turns, and the height difference of trajectory. From the specific data and the line graph, it can be seen that the method proposed in this paper was significantly better than the other algorithms in the number of iterations and the number of turns, reducing them by about 66.7%~87.5% and 80%~94.7%. Moreover, compared with other algorithms, the trajectory length was reduced by about 6%~23.3%. Furthermore, the MuL-ACO scheme did not choose a steep route to reduce the wear of the mobile robot in uneven terrain. The height difference of the trajectory was reduced by 15.9%~68.2%. As shown in Table 4, in Maps 1, 2, 3, and 6, although the running time of MuL-ACO was not the shortest, the gap between the running time of MuL-ACO and the shortest running time was no more than 0.3 s. In some complex maps (Map4 and Map5), the running time of MuL-ACO was even about 0.1 s~0.8 s less than other algorithms, which provides the possibility for trajectory planning in dynamic environments.

6.3. Simulation Experiment C

In this group of experiments, the Complex2 map of simulation experiment B was selected. Moreover, different starting points and ending points were randomly set to diversify the experimental results. Figure 11 shows the optimal trajectories planned by ACO, MH-ACO, MF-ACO, and MuL-ACO. Table 5 summarizes the qualitative comparison of the performance of algorithms.
In Figure 11a,d, compared with other algorithms, MuL-ACO found a flatter and smoother trajectory for the robot. Although MH-ACO and MF-ACO also tended to find flat trajectories, the planned trajectories were not smooth enough. In Figure 11b,c, the trajectories planned by different algorithms were concentrated in the same area with height. Moreover, the cost of bypassing this height area was high. The hybrid scheme proposed in this paper did not completely bypass the height area but planned a shorter and smoother trajectory with a small height difference.
As shown in Table 5, the MuL-ACO hybrid scheme proposed in this paper had great advantages in smoothing performance. In terms of the height difference of the trajectories, the trajectories planned by MuL-ACO tended to bypass uneven areas, and the total height difference of the trajectories was lower than that of ACO, MH-ACO, and MF-ACO. Further, compared to other algorithms, MuL-ACO had the least number of iterations and the shortest running time. The hybrid scheme proposed in this paper runs stably and plans trajectories with a high comprehensive quality.

6.4. Simulation Experiment D

In this section, the feasibility of the proposed scheme in dynamical environments, which consisted of 2 dynamic obstacles represented by red blocks and other static obstacles represented by black blocks, was tested. The starting point of the mobile robot was (3.5, 2.5), and the end point was set to (28.5, 28.5). The dynamic obstacles moved vertically down and right at a rate of 2 m per time step and moved a total of 6 steps. Figure 12 shows the process of MuL-ACO generating the optimal trajectory. In the initial trajectory generation stage, the planned trajectory avoided steep regions and dynamic obstacles in the environment and contained several redundant trajectory nodes. In particular, it can be seen that the planned trajectory actively bypassed the vertical downward moving obstacles and adjusted the trajectory to reach the end point. In the trajectory optimization stage, the length and smoothness of the initial trajectory were minimized by a mutual learning-based trajectory optimization algorithm. The specific length and smoothness of the initial trajectory and the optimized trajectory are given in Table 6. Clearly, the proposed scheme can stably plan desired trajectories in dynamic environments.

7. Conclusions

Aiming at the trajectory generation and optimization of mobile robots in an uneven environment, a hybrid scheme using mutual learning and adaptive ant colony optimization (MuL-ACO) was proposed in this paper. The initial trajectory was generated by an improved adaptive ant colony algorithm, and then a mutual learning-based trajectory optimization algorithm completed the trajectory optimization. The comprehensive heuristic function and the adaptive method based on the improved temperature reduction function greatly improved the performance of the ant colony algorithm. Another advantage is that the proposed scheme had a clear division of labor to stably provide high-quality feasible solutions.
Experiments conducted in uneven environments of different scenes and sizes modeled by 2D-H maps showed that the trajectory planned by MuL-ACO was superior to the other five algorithms in terms of smoothness, height difference, length, and algorithm convergence. Especially in more complex and larger maps, the MuL-ACO scheme was more adaptable and could stably plan a trajectory with a higher comprehensive quality. In the future, the proposed hybrid method may be further optimized and could be applied to more complex scenarios, such as multi-agent scenarios with multiple objects and social interaction scenarios.

Author Contributions

Conceptualization: F.Q.; methodology: F.Q. and W.Y.; software: F.Q., W.Y., W.L. and K.X.; validation: W.Y. and C.L.; formal analysis, F.Q. and W.L.; investigation, W.L. and F.Q.; resources, F.Q. and W.Y.; data curation, F.Q. and W.Y.; writing—original draft, F.Q.; writing—review and editing, F.Q. and W.Y.; visualization, F.Q.; supervision, F.Q. and C.L.; project administration, F.Q. and W.Y.; funding acquisition, F.Q. and W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation (grant nos. 61602529 and 61672539). This work was also supported by the Hunan Key Laboratory of Intelligent Logistics Technology (2019TP1015) and Scientific Research Project of Hunan Education Department (No. 17C1650).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACOadaptive ant colony optimization
MuL-ACOmutual learning and adaptive ant colony optimization
MH-ACOmultiple heuristics adaptive ant colony optimization
MF-ACOmulti-factor adaptive ant colony optimization
SAsimulated annealing
RGB-DRed Green Blue-Depth
APFartificial potential field method
GAgenetic algorithms
PSOparticle swarm optimization
GWOgrey wolf optimizer
IAACAimproved adaptive ant colony algorithm
MLTOMutual learning-based Trajectory Optimization Algorithm
SSAsparrow search algorithm

References

  1. Rubio, F.; Valero, F.; Llopis-Albert, C. A review of mobile robots: Concepts, methods, theoretical framework, and applications. Int. J. Adv. Robot. Syst. 2019, 16, 1–22. [Google Scholar] [CrossRef] [Green Version]
  2. Lindemann, R.A.; Bickler, D.B.; Harrington, B.D.; Ortiz, G.M.; Voothees, C.J. Mars exploration rover mobility development. IEEE Robot. Autom. Mag. 2006, 13, 72–82. [Google Scholar] [CrossRef]
  3. Ward, C.C.; Iagnemma, K. A dynamic-model-based wheel slip detector for mobile robots on outdoor terrain. IEEE Trans. Robot. 2008, 24, 821–831. [Google Scholar] [CrossRef]
  4. Ganganath, N.; Cheng, C.T.; Chi, K.T. Finding energy-efficient paths on uneven terrains. In Proceedings of the 2014 10th France-Japan/8th Europe-Asia Congress on Mecatronics (MECATRONICS2014-Tokyo), Tokyo, Japan, 27–29 November 2014; pp. 383–388. [Google Scholar]
  5. Mei, Y.; Lu, Y.H.; Hu, Y.C.; Lee, C.G. Energy-efficient motion planning for mobile robots. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), New Orleans, LA, USA, 26 April–1 May 2004; pp. 4344–4349. [Google Scholar]
  6. Park, K.; Kim, S.; Sohn, K. High-precision depth estimation with the 3d lidar and stereo fusion. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 2156–2163. [Google Scholar]
  7. Hara, Y.; Tomono, M. Moving object removal and surface mesh mapping for path planning on 3D terrain. Adv. Robot. 2020, 6, 375–387. [Google Scholar] [CrossRef]
  8. Shin, J.; Kwak, D.; Kwak, K. Model Predictive Path Planning for an Autonomous Ground Vehicle in Rough Terrain. Int. J. Control Autom. 2021, 6, 2224–2237. [Google Scholar] [CrossRef]
  9. Song, S.; Jo, S. Online inspection path planning for autonomous 3D modeling using a micro-aerial vehicle. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 6217–6224. [Google Scholar]
  10. Nam, T.H.; Shim, J.H.; Cho, Y.I. A 2.5 D map-based mobile robot localization via cooperation of aerial and ground robots. Sensors 2017, 17, 2730. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Jiang, G.; Yin, L.; Jin, S.; Tian, C.; Ma, X.; Ou, Y. A simultaneous localization and mapping (SLAM) framework for 2.5 D map building based on low-cost LiDAR and vision fusion. Appl. Sci. 2019, 9, 2105. [Google Scholar] [CrossRef] [Green Version]
  12. Asvadi, A.; Peixoto, P.; Nunes, U. Detection and tracking of moving objects using 2.5 d motion grids. In Proceedings of the 2015 IEEE 18th International Conference on Intelligent Transportation Systems, Gran Canaria, Spain, 15–18 September 2015; pp. 788–793. [Google Scholar]
  13. Patle, B.K.; Pandey, A.; Parhi, D.R.K.; Jagadeesh, A. A review: On path planning strategies for navigation of mobile robot. Def. Technol. 2019, 15, 582–606. [Google Scholar] [CrossRef]
  14. Barraquand, J.; Langlois, B.; Latombe, J.C. Numerical potential field techniques for robot path planning. IEEE Trans. Syst. Man Cybern. 1992, 22, 224–241. [Google Scholar] [CrossRef]
  15. Li, G.; Tong, S.; Cong, F.; Yamashita, A.; Asama, H. Improved artificial potential field-based simultaneous forward search method for robot path planning in complex environment. In Proceedings of the 2015 IEEE/SICE International Symposium on System Integration (SII), Nagoya, Japan, 11–13 December 2015; pp. 760–765. [Google Scholar]
  16. Zhang, G.L.; Hu, X.M.; Chai, J.F.; Zhao, L.; Yu, T. Summary of path planning algorithm and its application. Mod. Mach. 2011, 5, 85–90. [Google Scholar]
  17. Haldurai, L.; Madhubala, T.; Rajalakshmi, R. A study on genetic algorithm and its applications. Int. J. Comput. Sci. Eng. 2016, 4, 139. [Google Scholar]
  18. Syed, U.A.; Kunwar, F.; Iqbal, M. Guided Autowave Pulse Coupled Neural Network (GAPCNN) based real time path planning and an obstacle avoidance scheme for mobile robots. Robot. Auton. Syst. 2014, 62, 474–486. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Gong, D.W.; Zhang, J.H. Robot path planning in uncertain environment using multi-objective particle swarm optimization. Neurocomputing 2013, 103, 172–185. [Google Scholar] [CrossRef]
  20. Brand, M.; Masuda, M.; Wehner, N.; Yu, X.H. Ant colony optimization algorithm for robot path planning. In Proceedings of the 2010 International Conference on Computer Design and Applications, Qinhuangdao, China, 25–27 June 2010; pp. 3–436. [Google Scholar]
  21. Katoch, S.; Chauhan, S.S.; Kumar, V. A review on genetic algorithm: Past, present, and future. Multimedia Tools and Applications. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef]
  22. Li, W.; Meng, X.; Huang, Y.; Fu, Z.H. Multipopulation cooperative particle swarm optimization with a mixed mutation strategy. Inform. Sci. 2020, 529, 179–196. [Google Scholar] [CrossRef]
  23. Liu, X.; Zhang, D.; Zhang, T.; Zhang, J.; Wang, J. A new path plan method based on hybrid algorithm of reinforcement learning and particle swarm optimization. Eng. Comput. 2021, 39, 993–1019. [Google Scholar] [CrossRef]
  24. Jiao, Z.; Ma, K.; Rong, Y.; Wang, P.; Zhang, H.; Wang, S. A path planning method using adaptive polymorphic ant colony algorithm for smart wheelchairs. J. Comput. Sci-Neth. 2018, 25, 50–57. [Google Scholar] [CrossRef]
  25. Nie, Z.; Zhao, H. Research on robot path planning based on Dijkstra and Ant colony optimization. In Proceedings of the 2019 International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Shanghai, China, 21–24 November 2019; pp. 222–226. [Google Scholar]
  26. Chen, S.; Yang, J.; Li, Y.; Yang, J. Multiconstrained network intensive vehicle routing adaptive ant colony algorithm in the context of neural network analysis. Complexity 2017, 25, 1–9. [Google Scholar] [CrossRef] [Green Version]
  27. Li, L.I.; Hong, L.I.; Ningbo, S.H.A.N. Path planning based on improved ant colony algorithm with multiple inspired factor. Comput. Eng. Appl. 2019, 55, 219–225. [Google Scholar]
  28. Akka, K.; Khaber, F. Mobile robot path planning using an improved ant colony optimization. Int. J. Adv. Robot. Syst. 2018, 15, 1729881418774673. [Google Scholar] [CrossRef] [Green Version]
  29. Ning, J.; Zhang, Q.; Zhang, C.; Zhang, B. A best-path-updating information-guided ant colony optimization algorithm. Inform. Sci. 2018, 433, 142–162. [Google Scholar] [CrossRef]
  30. Chen, X.; Kong, Y.; Fang, X.; Wu, Q. A fast two-stage ACO algorithm for robotic path planning. Neural. Comput. Appl. 2013, 22, 313–319. [Google Scholar] [CrossRef]
  31. Liwei, Y.; Lixia, F.; Ning, G. Multi factor improved ant colony algorithm path planning. Comput. Integr. Manuf. Syst. (CIMS) 2021, 27, 1–18. [Google Scholar]
  32. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  33. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
Figure 1. The modelling process of the 2D-H map. (a) The 3D map of peak function. (b) The 2D-H map generated by the peak function and obstacles.
Figure 1. The modelling process of the 2D-H map. (a) The 3D map of peak function. (b) The 2D-H map generated by the peak function and obstacles.
Applsci 12 04629 g001
Figure 2. (a) The moving direction of the robot; (b) The planned trajectory limited by the height threshold.
Figure 2. (a) The moving direction of the robot; (b) The planned trajectory limited by the height threshold.
Applsci 12 04629 g002
Figure 3. Adaptive change of the pheromone volatilization factor.
Figure 3. Adaptive change of the pheromone volatilization factor.
Applsci 12 04629 g003
Figure 4. Convergence curves for PSO, GWO, ACO, and IAACA (μ = 33, μ = 50, μ = 90) on different benchmark functions. (a) Convergence performance of swarm intelligence algorithms on the Sphere function. (b) Convergence performance of swarm intelligence algorithms on the Rastrigin function. (c) Convergence performance of swarm intelligence algorithms on the Ackley function.
Figure 4. Convergence curves for PSO, GWO, ACO, and IAACA (μ = 33, μ = 50, μ = 90) on different benchmark functions. (a) Convergence performance of swarm intelligence algorithms on the Sphere function. (b) Convergence performance of swarm intelligence algorithms on the Rastrigin function. (c) Convergence performance of swarm intelligence algorithms on the Ackley function.
Applsci 12 04629 g004
Figure 5. The initial trajectory and the trajectory optimized by mutual learning.
Figure 5. The initial trajectory and the trajectory optimized by mutual learning.
Applsci 12 04629 g005
Figure 6. The mutual learning process of P 1 .
Figure 6. The mutual learning process of P 1 .
Applsci 12 04629 g006
Figure 7. Flow chart of the MuL-ACO.
Figure 7. Flow chart of the MuL-ACO.
Applsci 12 04629 g007
Figure 8. (af) Optimal trajectories generated by GA, SSA, and MuL-ACO on different 20 × 20 maps.
Figure 8. (af) Optimal trajectories generated by GA, SSA, and MuL-ACO on different 20 × 20 maps.
Applsci 12 04629 g008
Figure 9. The trajectories generated by ACO, MH-ACO, MF-ACO, and MuL-ACO. (a) Map. NO.1: X-type Map; (b) Map. NO.2: Z-type map; (c) Map. NO.3: Complex1 map; (d) Map. NO.4: Complex2 map; (e) Map. NO.5: Complex3 map; (f) Map. NO.6: Vortex map.
Figure 9. The trajectories generated by ACO, MH-ACO, MF-ACO, and MuL-ACO. (a) Map. NO.1: X-type Map; (b) Map. NO.2: Z-type map; (c) Map. NO.3: Complex1 map; (d) Map. NO.4: Complex2 map; (e) Map. NO.5: Complex3 map; (f) Map. NO.6: Vortex map.
Applsci 12 04629 g009
Figure 10. Comparison of trajectory performance. (a) length, (b) number of turns, (c) number of iterations, (d) the height difference of the trajectory.
Figure 10. Comparison of trajectory performance. (a) length, (b) number of turns, (c) number of iterations, (d) the height difference of the trajectory.
Applsci 12 04629 g010
Figure 11. (ad) The trajectories are generated by ACO, MH-ACO, MF-ACO, and MuL-ACO on the map with different starting points and ending points.
Figure 11. (ad) The trajectories are generated by ACO, MH-ACO, MF-ACO, and MuL-ACO on the map with different starting points and ending points.
Applsci 12 04629 g011
Figure 12. (af) Results of trajectory generation and optimization at different time steps (step = 1–6) in dynamic environments.
Figure 12. (af) Results of trajectory generation and optimization at different time steps (step = 1–6) in dynamic environments.
Applsci 12 04629 g012
Table 1. Description of benchmark functions.
Table 1. Description of benchmark functions.
FunctionsFunction ExpressionsRange D im f min
Sphere F 1 = i = 1 D x i 2 [−5.12, 5.12]100
Rastrigin F 2 = i = 1 D x i 2 10 cos 2 π x i + 10 [−5.12, 5.12]100
Ackley F 3 = 20 exp 0.02 1 D i = 1 D x i exp 1 D i = 1 D cos 2 π x i + 20 + e [−5.12, 5.12]100
Table 2. Simulation results of GA, SSA, and MuL-ACO.
Table 2. Simulation results of GA, SSA, and MuL-ACO.
NameMap NO.LengthIterationsTurnsHeight DifferenceTime (s)
GAa28.732105.69.2
b29.73895.412.5
c28.53784.210.7
d28.13155.310.4
e28.63976.712.6
f29.03464.19.5
SSAa29.62894.30.7
b29.92595.70.9
c30.72673.10.5
d31.43762.80.8
e28.921107.20.6
f29.32367.50.6
MuL-ACOa32.4750.91.4
b29.8661.21.3
c28.0820.31.5
d27.9930.51.7
e28.7630.41.3
f28.2722.11.6
Table 3. Simulation environment description.
Table 3. Simulation environment description.
Map NO.NameStart PointEnd Point
1X-type(0.5, 0.5)(25.0, 23.0)
2Z-type(1.0, 24.0)(25.0, 10.0)
3Complex1(0.5, 0.5)(26.0, 23.0)
4Complex2(0.5, 0.5)(29.0, 24.0)
5Complex3(1.0, 0.5)(29.0, 27.0)
6Vortex(28.0, 20.0)(13.0, 15.0)
Table 4. Simulation results of ACO, MH-ACO, MF-ACO, and MuL-ACO.
Table 4. Simulation results of ACO, MH-ACO, MF-ACO, and MuL-ACO.
NameMap NO.LengthIterationsTurnsHeight DifferenceTime (s)
ACO160.8040387.592.02
267.41392516.712.73
341.238227.361.41
443.6032176.481.63
543.030147.451.71
6——————————
MH-ACO156.8227152.292.30
266.3021139.312.53
344.2128146.721.54
443.402585.101.72
543.6023121.551.74
678.4019318.152.57
MF-ACO154.2715103.441.93
274.1012149.982.87
344.5723155.541.96
443.411893.332.28
546.801374.382.39
673.6216226.932.54
MuL-ACO146.38621.512.11
260.21565.342.65
342.88592.181.61
439.70720.421.52
543.10831.561.55
655.93975.882.83
Table 5. Simulation results of ACO, MH-ACO, MF-ACO, and MuL-ACO on the map with different starting points and ending points.
Table 5. Simulation results of ACO, MH-ACO, MF-ACO, and MuL-ACO on the map with different starting points and ending points.
NameMap
(with Different Start and End Points)
LengthIterationsTurnsHeight
Difference
Time (s)
ACOa34.8037116.811.51
b35.8025167.441.40
c31.6223123.631.53
d38.60152111.021.52
MH-ACOa41.802350.341.63
b35.201751.491.53
c32.801482.651.55
d41.001180.381.66
MF-ACOa36.601772.402.20
b36.201362.072.10
c32.20943.232.24
d40.201044.332.13
MuL-ACOa40.091130.391.38
b34.84931.491.12
c30.891033.041.24
d39.42640.381.50
Table 6. Simulation results of (un)MH-ACO and MuL-ACO.
Table 6. Simulation results of (un)MH-ACO and MuL-ACO.
NameMap NO.LengthTurns
(un)MuL-ACOa41.67
b42.09
c40.811
d40.28
e40.87
f42.66
MuL-ACOa39.22
b39.13
c38.64
d38.33
e39.52
f39.32
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qu, F.; Yu, W.; Xiao, K.; Liu, C.; Liu, W. Trajectory Generation and Optimization Using the Mutual Learning and Adaptive Ant Colony Algorithm in Uneven Environments. Appl. Sci. 2022, 12, 4629. https://doi.org/10.3390/app12094629

AMA Style

Qu F, Yu W, Xiao K, Liu C, Liu W. Trajectory Generation and Optimization Using the Mutual Learning and Adaptive Ant Colony Algorithm in Uneven Environments. Applied Sciences. 2022; 12(9):4629. https://doi.org/10.3390/app12094629

Chicago/Turabian Style

Qu, Feng, Wentao Yu, Kui Xiao, Chaofan Liu, and Weirong Liu. 2022. "Trajectory Generation and Optimization Using the Mutual Learning and Adaptive Ant Colony Algorithm in Uneven Environments" Applied Sciences 12, no. 9: 4629. https://doi.org/10.3390/app12094629

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop