Next Article in Journal
Combined Polarization/Magnetic Modulation of a Transverse NMR Gyroscope
Previous Article in Journal
An Evidence Theoretic Approach for Traffic Signal Intrusion Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Biologically Inspired Complete Coverage Path Planning Algorithm Based on Q-Learning

1
CAS Key Laboratory of On-Orbit Manufacturing and Integration for Space Optics System, Chinese Academy of Sciences, Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
Research Center for Materials and Optoelectronics, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4647; https://doi.org/10.3390/s23104647
Submission received: 16 March 2023 / Revised: 24 April 2023 / Accepted: 28 April 2023 / Published: 11 May 2023
(This article belongs to the Section Sensors and Robotics)

Abstract

:
Complete coverage path planning requires that the mobile robot traverse all reachable positions in the environmental map. Aiming at the problems of local optimal path and high path coverage ratio in the complete coverage path planning of the traditional biologically inspired neural network algorithm, a complete coverage path planning algorithm based on Q-learning is proposed. The global environment information is introduced by the reinforcement learning method in the proposed algorithm. In addition, the Q-learning method is used for path planning at the positions where the accessible path points are changed, which optimizes the path planning strategy of the original algorithm near these obstacles. Simulation results show that the algorithm can automatically generate an orderly path in the environmental map, and achieve 100% coverage with a lower path repetition ratio.

1. Introduction

Along with the unceasing progress of science and technology, all kinds of mobile robots have been imagined, and these robots are applied to many fields such as room cleaning [1,2,3], exploration [4], search-and-rescue [5,6], and guard patrol [7]. As one of the most important links with robot technology, path planning has become increasingly important to researchers. Path planning can be divided into two categories: the first is point-to-point path planning [8,9], and the other is complete coverage path planning [10]. Point-to-point path planning needs the robot to plan a path line from the start point to the finish point which does not interfere with the obstacles in the environmental map. Complete coverage path planning needs the robot to plan a path in the environmental map which includes all reachable positions and avoids repeated paths. Traditional complete coverage path planning algorithms are mainly as follows: the stochastic and ergodic algorithm [11], the boustrophedon coverage algorithm [12], the interior screw algorithm [13], and the domain decomposition algorithm [14]. These algorithms are easy to operate and the calculation speed is fast; however, when it comes to practical application, these algorithms have poor efficiency and a high coverage repetition ratio for the planned path.
With the rapid development of artificial intelligence-related technologies, many intelligent algorithms are used for the complete coverage path planning domain [15], such as the genetic algorithm [16], the ant colony algorithm [17], and the biologically inspired neural network algorithm. The biologically inspired neural network algorithm has the following advantages: there is no need to train the model and no need to anticipate environmental information; in addition, it operates with fast computing speed. Therefore, it has been widely studied and applied by many scholars. Yang et al. [18] applied the biologically inspired neural network algorithm to the complete coverage path planning domain for the first time. In their algorithm, the map is shown as a grid map, and each grid is considered as a neuron. The adjacent neurons are connected, and the robot’s mobile path is chosen by calculating the neuronal activity value. Zhu et al. [19] improved the biologically inspired neural network algorithm by presenting the Glasius bionic neural network algorithm (GBNN), which reduced the calculating time of algorithm path planning. Lou et al. [20] presented a complete coverage neural network (CCNN) algorithm, which simplified the calculation process of neural activity. It also combined a modified A* algorithm, which effectively lowered the number of path-turning times, and improved the coverage efficiency. Zhang Zhiyuan et al. [21] brought in the domain neuron status criterion, which increased the biologically inspired path planning efficiency near isolated island obstacles. Qian Jinwei et al. [22] combined the interior screw search and the biologically inspired neural network algorithm, which reduced the calculating time of the path repetition ratio. Curiac et al. [23] used the 2D Arnold cat map as a primary positional entropy source, and proposed a new bio-inspired method to generate deceptive paths for mobile robots to complete patrolling missions under the threat of confrontation. Moysis et al. [24] proposed a simple three-parameter one-dimensional chaotic map, and the proposed pseudo-random bit generator was applied to the problem of chaotic path planning for a UAV exploring a 3D area. Moysis et al. [25] proposed a simple, short, and efficient chaotic path planning algorithm for autonomous mobile robots to cover a given terrain using chaotic and unpredictable motion.
On the other hand, the rapid development of reinforcement learning and deep learning brings a whole new solution to path planning. Ai et al. [26] presented a complete coverage path planning algorithm based on reinforcement learning, which modeled optimized path choice by a reward function with multiclass constraints. Dong Jixin [27] presented a path planning algorithm based on deep learning, which modeled a perceived grid map by coiling base, and he set up an intelligent grid based on resnet-18 grid structure, then accomplished complete coverage on static and dynamic maps by corresponding training. The reinforcement learning method possesses a strong ability for decision making and can find the global optimal solution; however, it costs much more time during the learning process than a generic algorithm, so its application has not been widely used.
Among all aforementioned algorithms, the complete coverage path planning algorithm has the fastest calculating speed, but the path planning repetition ratio is high in complex environments; intelligent algorithms considering environmental information can plan orderly paths, but have the problem of long calculating times. In response to the issue and research contents, this paper presents a biologically inspired neural network algorithm based on Q-learning, which proceeds with path planning based on the biologically inspired neural network algorithm, and when it comes to the position change of a reachable path point, processes path optimization by adopting Q-learning to avoid falling into a locally optimal solution, thus reducing the path repetition ratio, while still ensuring complete coverage, achieving the desired path planning results. The rest of this paper is organized as follows. In Section 2, the complete coverage path planning algorithm based on a biologically inspired neural network is described in detail. In Section 3, the improved algorithm based on Q-learning is introduced. Simulation results and discussions in different environmental maps are presented in Section 4. Finally, the conclusion is given in Section 5.

2. Biologically Inspired Neural Network Algorithm

The complete coverage path planning issue can be described as follows: on an environmental map, the mobile robot sets off at the starting point, avoids obstacles, passes by all the reachable position points in an orderly way on the environmental map, and reduces path repetition as much as possible. The biologically inspired neural network algorithm has many advantages such as the following: simple calculation, no need for model training, and the mobile robot can plan a 100% coverage ordered path under the restricted condition of no motion template or energy loss function; therefore, it is widely used in the indoor robot complete coverage path planning domain.

2.1. Biologically Inspired Neural Network Model

Physicists Hodgkin and Huxley developed the famous biological nerve cell membrane circuit model [28] to explain the signal transmission issue in nervous systems, where the dynamic membrane voltage V m can be expressed as the following formula:
C m d V m d t = E p + V m g p + E N a V m g N a E k + V m g K
where Cm is the membrane capacitance, Ek, Ena, Ep, gk, gna, and gp are electric potentials and the electric conductance of potassium ions, sodion, and the passive leakage current in cell membranes, respectively. Grossberg made a summing based on Equation (1), and set up the neural dynamic network model, and created the famous shunting equation:
d x i d t = A x i + B x i S i e ( t ) D + x i S i i ( t )
where x i can be considered as the neural activity value of the neuron i ; A, B, and D are the attenuation ratio, the upper bound, and the lower bound of the neural activity value, respectively; S i e and S i i are the external stimulus and the inhibiting input of the neuron, respectively. The fractional flow Equation (2) is widely used in the domain of visual perception and motion control.
Yang et al. [29] set up a biologically inspired neural network based on Equation (2), and applied it to the path planning domain, which achieved good results. In a neural network, the size of the neuron activity value is decided by the following formula:
d x i d t = A i + B x i I i + + j = 1 k ω i j [ x j ] + D + x i I i
where x i and x j are the neuron activity values of neurons i and j , k is the amount of all the adjacent neurons of neuron i , and the representative content of A , B , and D are the same as in Equation (2). I i + + j = 1 k ω i j [ x j ] + and I i are the external stimulus and the inhibiting input of the neuron. The functions f + and f are defined as f + = max f , 0 and f = max f , 0 , and ω i j is the weight connection of neurons i and j , which can be defined as follows:
ω i j = μ | i j | 0 < | i j | < r 0 | i j | r
where i j is the Descartes distance of the neurons on the grid map, and μ is a positive constant. The value of r decides the connection range of the neurons, as shown in Figure 1, when the value of r = 2 , and the neuron i connects to 8 neurons around i . The external input I i in Equation (3) can be defined as follows:
I i = E target   point   of   neuron E obstacle   point   of   neuron 0 other   point   of   neuron
where E is a large number positive constant, to express a large external input.
The above Equation (3) can ensure that the positive neuronal activity can be spread into the entire neural network by the connection of neurons, and the negative neuronal activity can be limited to the nearby obstacles to block signal transmission. Therefore, using the fractional flow formula to define the model, the mobile robot can automatically choose the next moment’s moving direction by the neuron activity value of the neural network, then complete the path planning. The above neural network model can be described in Figure 1.

2.2. Biologically Inspired Complete Coverage Path Planning

Yang et al. [8] proceeded with path point optimization selection on the above biologically inspired neural network model, and applied it to the complete coverage path planning domain. In order to generate an ordered and low repetition ratio path, the deciding path points can be defined as the following equation:
P n     x p n = max x j + c y j ,   j = 1 , 2 , , k
where P c is the current position of the mobile robot, P n is the next moment’s position of the mobile robot, c is a positive constant, y j is the function to define the steering angle of the mobile robot, which can be determined by the following Equation:
y j = 1 Δ θ j π
Δ θ j = Δ θ x π Δ θ r 2 π x > π
Δ θ r = θ j θ c = a t a n 2 y p j y p c , x p j x p c a t a n 2 y p c y p p , x p c x p p
where P p is the next moment’s position of the mobile robot; Δ θ j 0 , π is the angle variation of the next moment’s moving direction of the mobile robot; a t a n 2 y a , x a π , π is the four quadrant arc-tangent functions to describe the angle variation of the moving direction.
According to the above path point selection method, the biologically inspired complete coverage path planning process is as follows:
(a)
Build a neural network model correlated with environmental information, and set up external excitation and inhibiting input according to obstacles in the map and the positions of reachable points;
(b)
Iteratively calculate each neuron’s activity value according to the neural network model and external input;
(c)
The mobile robot selects the next moment’s path point according to the activity value of the neuron adjacent to the current position and steering angle variation;
(d)
When the robot falls into a dead zone, which means that around its current position are obstacles and the path point has been passed, the robot calculates the activity value of the neuron grid adjacent to the current position; if the values are all less than the value of the current position, the robot remains still until there is one path point’s activity value greater than the current position value; then the robot moves into the position of the next moment’s path point;
(e)
The mobile robot considers the next moment’s path point as the current position, then estimates if it will accomplish the complete coverage of the environmental map. If the answer is yes, then it is the end of the algorithm; if the answer is no, then go back to step (b).
According to the above content, a flow chart of the biologically inspired complete coverage path planning algorithm based on Q-learning is shown as follows (Figure 2):

3. Improved Algorithm Based on Q-Learning

The biologically inspired neural network algorithm has the merit of simple calculation; however, the repetition ratio of planned paths is very high in an environment with obstacles distributed in complexity; therefore, this paper presents the improved algorithm based on Q-learning, which optimizes the path combined with the environmental map.

3.1. Issue Description

According to the above algorithm principle, build a 15 × 15 grid map corresponding to the neural network model. The mobile robot starts to plan a path at the starting point (2,2); in order to generate a more ordered path, the neuron of the neural network model only connects to four neurons whose directions are up, down, left, and right, which means the mobile robot can only move towards four directions. Parts of the planned path figure generated by the algorithm are shown in Figure 3, where the black rectangle represents obstacles, and the blue directed line segment represents the planned path.
As shown in Figure 3, the above biologically inspired neural network complete coverage algorithm can avoid obstacles, and moves automatically to uncovered position points; however, the path selection strategy is decided by Equation (6), and therefore, in most cases, the path selected by the robot is only extended in the direction of the last moment, until it meets obstacles or a planned path. The above moving strategy does not consider feature information of the environmental map, and it cannot adjust direction automatically according to the position of obstacles, so the generated path is unordered and the repetition ratio is high; when it comes to a more complex environment, this situation is more distinct. For that, this paper brings in Q-learning to optimize the selection on path direction, to reduce the repetition ratio of the algorithm model.

3.2. Principles of the Q-Learning Algorithm

Q-learning is a classic no model reinforcement learning algorithm, which is based on the theory of Markovian decision processes. Markovian decision processes can be defined by a tetrad S , A , P , R , where S represents discrete-time state space, A represents discrete-time motion space, P represents the probability function of state transition, R is the reward function, which represents the reward of state transition after the intellectual body takes an action. The target of the Markovian decision is to find an appropriate strategy, to maximize the cumulative mathematical expectation of the reward function. According to Behrman’s optimal criterion, the optimal index of the optimal strategy is as follows:
V * x = max a A R s , a + γ s S P s , a s V s
There is a state action function, Q function, in Q-learning for evaluation. The function value of Q s , a is the maximum cumulative reward reduced value after the motion a is executed from state S . The optimum strategy of this moment can be defined by the Q function, and there is no need to consider the state transition probability P , so by selecting the motion which maximizes the Q function from all the motions a in the state S , then the globally optimal strategy π can be obtained. The Q function iterative formula based on Behrman criterion is as follows:
Q s t , a = Q s t , a + α r t + γ m a x Q s t + 1 , a Q s t , a
where α 0 , 1 is the learning rate, r t is the rewards the intelligent body receives after performing the action, and γ 0 , 1 is the discount factor.
During the learning process, Q-learning sets up a Q form whose purpose is to store the Q value that corresponds to all the state actions of the intelligent body. The common action selection strategy of Q-learning is the ε greedy strategy. By setting the greedy factor ε 0 , 1 , the intelligent body will select the optimal action based on the Q form when the motion selection has ε probability, and will select a random action when the motion selection has 1 ε probability, which then gives the intelligent body the ability to explore actively. Through repeated learning, the Q form comes to convergence, and then the intelligent body can obtain each state’s optimal action through the Q form, which is the output of the optimal policy. The algorithm flow chart of Q-learning is shown in Figure 4.

3.3. Algorithm Improvements

Reinforcement learning has a strong ability for learning and decision making; however, applying a reinforcement learning algorithm on complete coverage path planning will create a problem: the convergence speed of the function Q is very slow, and sometimes it does not converge due to big environmental maps and sparse learning rewards, so entirely applying the Q-learning algorithm to a global map will need to be improved. The efficiency of the original biologically inspired neural network algorithm is high; however, it does not consider map features near obstacles or around the planned path, so applying the Q-learning algorithm on the area near obstacles and proceeding with strategy optimization can improve efficiency greatly. In order to increase the ratio of convergence of Q-learning, only apply Q-learning when the mobile robot detects that the reachable position points are increasing, and apply the biologically inspired neural network algorithm on the other areas where there is no change in the reachable position points, which can plan the path rapidly and orderly, as shown in Figure 5.
In Figure 5a, the robot in position 6 has only one reachable point, 7, and when the robot reaches position 7, the reachable position points increase to two points, 8 and 11, so this is the moment to apply Q-learning to the path planning. In Figure 5b, the robot in position 6 has two reachable points, 7 and 10, and when the robot reaches position 7, the reachable position points increase to three points, 3, 8, and 11, so this is the moment to apply Q-learning to the path planning. In contrast, in Figure 5c,d, during the mobile robot’s motor process, the number of reachable position points does not increase, so this is the moment to apply the biologically inspired neural network to the path planning. Figure 5 shows the situations that might happen during path planning, and the situations are judged by the number of reachable position points; if the number is increased, then apply Q-learning to the path planning.
In order to reduce algorithm complexity and generate more ordered paths, make a mobile robot that can only move towards up, down, left, and right (four directions) for reinforcement learning and the biologically inspired neural network algorithm. In order to accelerate the Q form convergence of this algorithm, the mobile robot is encouraged to change moving direction according to the position of obstacles when the reachable position points increase. The setting of the greedy strategy of the Q-learning process is shown as follows, where ε 0 , 1 , σ 0 , 1 :
π s t = a r g m a x   Q s t , a       σ ε a A       ε < σ ε + 2 × 1 ε 3 a B       ε + 1 ε 3 < σ
When the mobile robot chooses a moving direction, determine an exploring factor ε . The robot selects an optimal action in Q form based on the probability of the exploring factor ε . Select the action A with a different moving direction from the last moment’s action based on the probability of 2 × 1 ε 3 and randomly select the action B based on the probability of 1 ε 3 . At the beginning of learning, the Q form has not come to convergence, and the algorithm tends to explore the environment, which means selecting a smaller ε . After a period of learning, the algorithm tends to make use of the environment, and the number of ε should be set for a large number, to ensure finding the global optimal solution.
According to the above content, a flow chart of the biologically inspired complete coverage path planning algorithm based on Q-learning is shown as follows (Figure 6):
Specific steps of the algorithm according to the above algorithm flow are as follows:
(a)
Build an environmental map, and then initialize the parameters of the biologically inspired neural network and Q-learning;
(b)
Estimate if the reachable position points number of the mobile robot is increasing; if it is not increasing, proceed to step (c); if it is increasing, proceed to step (d).
(c)
Calculate each neuron’s activity value of the neural network according to the obstacle distribution in the obstacle environmental map and the planned path based on Equation (3), and then the robot moves according to Equation (6); proceed to step (e);
(d)
Select the next mobile strategy according to Q-learning’s greedy algorithm, proceed to moving and update Q form, and proceed to step (e);
(e)
Estimate if complete coverage path planning will be accomplished or not; if not, go back to step (b), and continue to plan paths; if yes, proceed to step (f);
(f)
Estimate if the Q form has been converged or not; if not, go back to step (a) and proceed with the next learning; if yes, then complete the path planning task, and output the path.

4. Simulation Experiment Analysis

In order to verify the feasibility and effectiveness of the biologically inspired complete coverage path planning algorithm based on Q-learning, proceed to the simulation experiment of mobile robot path planning on an indoor environmental map.

4.1. Setting of Simulation Conditions

Build a grid map according to the size of the environmental map and obstacle distribution. Set the starting point of the mobile robot. If the robot reaches all the coverage grid points, it is considered as having accomplished the complete coverage path planning.
Set partial parameters of the biologically inspired neural network as follows: in Equation (3), A = 100 , B = 1 , D = 1 , μ = 0.8 , r = 1 , output E = 100 , and in Equation (6), weighting c = 0.1 . Set partial parameters of Q-learning as follows: set four basic motions in the mobile robot motion space which are up, down, left, and right; and in Equation (11), α = 0.5 , γ = 0.7 . The settings of the reinforcement learning process reward are as follows: when the intelligent agent reaches an uncovered position point, the reward value is +3; when the intelligent agent reaches a covered position point, the reward value is −2; when the intelligent agent hits obstacles, the reward value is −10, and the planning task is considered a failure; when the intelligent agent’s moving steps are greater than two times the total grid number, the reward value is −15, and the planning task is considered a failure; when the intelligent agent reaches all the coverage position points, the reward value is +15, and the planning task is considered successful.

4.2. Indoor Map Simulation Experiment

Build a size 15 × 15 grid map according to an indoor environment, shown in Figure 7. The obstacle position information in the grid map is known, shown using a black rectangle. The mobile robot sets off at the starting point (2,2). The planned path is shown using the blue directed line segment. When there are no changes at the mobile robot’s reachable position point, the robot plans a path and moves using the biologically inspired neural network, and the mobile robot sheers off and avoids obstacles according to the neuron activity value corresponding to the grid point. When the reachable position point changes, adopt the Q-learning algorithm to choose a more reasonable path point, and sheer off automatically based on the obstacle map information, and the repetition ratio of the complete coverage path planning is reduced. On the same condition of the same experiment map and starting point, this paper compares the biologically inspired complete coverage path planning algorithm based on Q-learning with the traditional boustrophedon coverage algorithm [4] as shown in Figure 8 and the secondary region partitioning algorithm based on cellular automaton theory [30] as shown in Figure 9; the repetition ratio of path planning, the coverage ratio of path planning, and the number of sheering off times are shown in Table 1. According to the comparison results, on the same conditions, the algorithm proposed in this paper can achieve a 100% coverage ratio, and change path directions automatically based on the position of obstacles; although the sheering times are increased, the path repetition ratio is reduced effectively.
The above simulation results show that the biologically inspired complete coverage path planning algorithm based on Q-learning is very effective, and it can plan a better path compared with the traditional boustrophedon coverage algorithm and the secondary region partitioning algorithm. Furthermore, this paper conducts a simulation experiment on a bigger size indoor environmental map, shown in Figure 10. The sizes of the grid are 30 × 20 , and there are more types of obstacles in the environmental map, such as obstacles near the edge of the map and island-style obstacles. The simulation parameters are set to be the same as in the above simulation process, and the mobile robot starts at the position point (2,2); comparisons with the traditional boustrophedon coverage algorithm as shown in Figure 11 and the secondary region partitioning algorithm as shown in Figure 12 were conducted, and the results are shown in Table 2. The algorithm proposed in this paper can plan a complete coverage path equally with a lower repetition ratio on a bigger size indoor environmental map.
The simulation results show that the improved algorithm is different from the traditional complete coverage path planning algorithms, which show that the improved algorithm proposed in this paper can change planned path direction actively according to the distribution of obstacles on the map. Although the path generated by the improved algorithm has more turns, the path repetition rate is lower, which reduces the energy loss of the mobile robot at work. The improved algorithm has the characteristics of the high computational efficiency of the biologically inspired neural network, and the Q-learning algorithm is introduced to optimize the turning strategy of the path near the obstacles, so that the planned path is more orderly.

5. Conclusions

This paper presents a biologically inspired complete coverage path planning algorithm based on Q-learning, which uses Q-learning in areas near obstacles to improve the algorithm of path planning strategy based on the original biologically inspired neural network algorithm. The improved algorithm solves two problems: one is the complex paths and high repetition coverage ratio of the biologically inspired neural network algorithm; the other is avoiding the low computational efficiency of using only Q-learning on complete coverage path planning. The results indicate that the improved algorithm can generate a complete coverage path in an indoor environment without a motion template or energy loss function, and while reducing the path repetition ratio effectively.
This paper conducts research on robot complete coverage path planning in a simpler indoor environment; however, in actual applications, the environment is more complex in general. The Q-learning algorithm tends to find the globally optimal path, so the Q form converges very slowly if there are no constraints. Therefore, there are many involved possibilities during the complete coverage process, and the earlier reinforcement learning exploration stage costs much time, so the calculation speed is very slow and convergence is very difficult in complex and large-scale scenarios. Therefore, the next research content is to expand and improve the algorithm. For example, adding a number of scene restrictions on Q-learning, and optimizing the Q form initial value according to the data on the neuron activity value, may allow this algorithm to achieve complete coverage path planning under the conditions of complex environments such as multi-obstacle scenarios and large-scale scenarios.

Author Contributions

Conceptualization, X.T.; methodology, X.T.; software, L.H.; data curation, L.H.; writing—original draft preparation, X.T.; writing—review and editing, X.T.; visualization, H.G.; project administration, Q.W.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All relevant data are within the paper.

Acknowledgments

Thanks to Junlin Ma (CIOMP). He has provided many helpful suggestions to this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pathmakumar, T.; Muthugala, M.A.V.J.; Samarakoon, S.M.B.P.; Gómez, B.F.; Elara, M.R. A Novel Path Planning Strategy for a Cleaning Audit Robot Using Geometrical Features and Swarm Algorithms. Sensors 2022, 22, 5317. [Google Scholar] [CrossRef] [PubMed]
  2. Miao, X.; Lee, H.-S.; Kang, B.-Y. Multi-Cleaning Robots Using Cleaning Distribution Method Based on Map Decomposition in Large Environments. IEEE Access 2020, 8, 97873–97889. [Google Scholar] [CrossRef]
  3. Kong, S.; Wu, Z.; Qiu, C.; Tian, M.; Yu, J. An FM*-Based Comprehensive Path Planning System for Robotic Floating Garbage Cleaning. IEEE Trans. Intell. Transp. Syst. 2022, 23, 23821–23830. [Google Scholar] [CrossRef]
  4. Xu, Y.; Zhang, Z.; Yu, J.; Shen, Y.; Wang, Y. A Framework to Co-Optimize Robot Exploration and Task Planning in Unknown Environments. IEEE Robot. Autom. Lett. 2022, 7, 12283–12290. [Google Scholar] [CrossRef]
  5. Liu, Y.; Chen, Z.; Li, Y.; Lu, M.; Chen, C.; Zhang, X. Robot Search Path Planning Method Based on Prioritized Deep Reinforcement Learning. Int. J. Control. Autom. Syst. 2022, 20, 2669–2680. [Google Scholar] [CrossRef]
  6. Li, L.; Shi, D.; Jin, S.; Yang, S.; Zhou, C.; Lian, Y.; Liu, H. Exact and Heuristic Multi-Robot Dubins Coverage Path Planning for Known Environments. Sensors 2023, 23, 2560. [Google Scholar] [CrossRef] [PubMed]
  7. Lu, Y.-H.; Juang, J.-G. Application of Differential Global Positioning System and Path Planning to Robot Outdoor Patrol. Sens. Mater. 2018, 30, 1643–1654. [Google Scholar] [CrossRef]
  8. Lin, H.X.; Xiang, D.; Ouyang, J.; Xiaodong, L. Review of path planning algorithms for mobile robots. Comput. Eng. Appl. 2021, 57, 38–48. [Google Scholar]
  9. Alqarni, M.A.; Saleem, S.; Alkatheiri, M.S.; Chauhdary, S.H. Optimized path planning of drones for efficient logistics using turning point with evolutionary techniques. J. Electron. Imaging 2022, 31, 061819. [Google Scholar] [CrossRef]
  10. Zhu, D.; Zhu, T.; Yan, M. Multi-AUV complete coverage path planning based on improved neural network. J. Syst. Simul. 2020, 32, 1505–1514. [Google Scholar]
  11. Liu, P.; Sun, J.; Wang, C. Research on random coverage path planning of chaotic mobile robots. Control. Eng. China 2021, 28, 646–655. [Google Scholar]
  12. Viet, H.H.; Dang, V.H.; Laskar, M.N.; Chung, T. BA*: An online complete coverage algorithm for cleaning robots. Appl. Intell. 2013, 39, 217–235. [Google Scholar] [CrossRef]
  13. Zheng, R.; Liu, Z.P.; Yi, B.; Yang, Y. Adaptive spiral path planning method for additive manufacturing. Comput. Integr. Manuf. Syst. 2021, 27, 2016–2022. [Google Scholar]
  14. Zhou, L.N.; Wang, Y.; Zhang, X.; Yang, C.Y. Complete coverage path planning of mobile robot on abandoned mine land. Chin. J. Eng. 2020, 42, 1220–1228. [Google Scholar]
  15. Yu, M. A solution of TSP based on the ant colony algorithm improved by particle swarm optimization. Discret. Contin. Dyn. Syst.-S 2018, 12, 979–987. [Google Scholar] [CrossRef]
  16. Shang, G.; Liu, G.; Zhu, P. Complete Coverage Path Planning for Horticultural Electric Tractors Based on an Improved Genetic Algorithm. J. Appl. Sci. Eng. 2021, 24, 447–456. [Google Scholar]
  17. Ji, H.; Hu, H.; Peng, X. Multi-Underwater Gliders Coverage Path Planning Based on Ant Colony Optimization. Electronics 2022, 11, 3021. [Google Scholar] [CrossRef]
  18. Yang, S.X.; Luo, C. A neural network approach to complete coverage path planning. IEEE Trans. Syst. Man Cybern. Part B—Cybern. 2004, 34, 718. [Google Scholar] [CrossRef]
  19. Zhu, D.; Tian, C.; Sun, B.; Luo, C. Complete coverage path planning of autonomous underwater vehicle based on GBNN algorithm. J. Intell. Robot. Syst. 2018, 94, 237–249. [Google Scholar] [CrossRef]
  20. Xu, P.F.; Ding, Y.X.; Luo, J.C. Complete coverage path planning of an unmanned surface vehicle based on a complete coverage neural network algorithm. J. Mar. Sci. Eng. 2021, 9, 1163. [Google Scholar] [CrossRef]
  21. Zhang, Z.Y.; Zhao, X.; Jin, Y. Improvement of traversal path planning algorithm of cleaning robot based on biologically inspired neural network. J. Light Ind. 2018, 33, 73–78+85. [Google Scholar]
  22. Qian, J.W.; Dai, X.Q.; Gao, H.B.; Zhu, Y.Q. Path planning algorithm for biological incentive traversal based on internal spiral search. Comput. Simul. 2021, 38, 339–343+394. [Google Scholar]
  23. Curiac, D.I.; Banias, O.; Volosencu, C.; Curiac, C.D. Novel Bioinspired Approach Based on Chaotic Dynamics for Robot Patrolling Missions with Adversaries. Entropy 2018, 20, 378. [Google Scholar] [CrossRef] [PubMed]
  24. Moysis, L.; Rajagopal, K.; Tutueva, A.V.; Volos, C.; Teka, B.; Butusov, D.N. Chaotic Path Planning for 3D Area Coverage Using a Pseudo-Random Bit Generator from a 1D Chaotic Map. Mathematics 2021, 9, 1821. [Google Scholar] [CrossRef]
  25. Moysis, L.; Petavratzis, E.; Volos, C.; Nistazakis, H.; Stouboulos, I. A chaotic path planning generator based on logistic map and modulo tactics. Robot. Auton. Syst. 2020, 124, 103377. [Google Scholar] [CrossRef]
  26. Ai, B.; Li, B.; Gao, S.; Xu, J.; Shang, H. An intelligent decision algorithm for the generation of maritime search and rescue emergency response plans. IEEE Access 2019, 7, 155835–155850. [Google Scholar] [CrossRef]
  27. Dong, J.X. Area coverage path planning of UAV based on deep reinforcement learning. Ind. Control. Comput. 2021, 34, 80–82. [Google Scholar]
  28. Hodgkin, A.L.; Huxley, A.F. A quantitative description of membrane current and its application to conduction and excitation in nerve. Bull. Math. Biol. 1989, 52, 25–71. [Google Scholar] [CrossRef]
  29. Yang, S.X.; Meng, M. Neural network approaches to dynamic collision-free trajectory generation. IEEE Trans. Syst. Man Cybern. Part B—Cybern. 2001, 31, 302–318. [Google Scholar] [CrossRef]
  30. Jiang, L.; Zhang, Y.F.; Ma, X.Z.; Zhu, J.Y.; Lei, B. Complete coverage path planning based on secondary area division. J. Harbin Eng. Univ. 2022, 43, 1483–1490. [Google Scholar]
Figure 1. The neural network’s model.
Figure 1. The neural network’s model.
Sensors 23 04647 g001
Figure 2. Flow chart of biologically inspired complete coverage path planning algorithm.
Figure 2. Flow chart of biologically inspired complete coverage path planning algorithm.
Sensors 23 04647 g002
Figure 3. Part of the path planned by the biologically inspired neural network algorithm.
Figure 3. Part of the path planned by the biologically inspired neural network algorithm.
Sensors 23 04647 g003
Figure 4. Flow chart of the Q-learning algorithm.
Figure 4. Flow chart of the Q-learning algorithm.
Sensors 23 04647 g004
Figure 5. Variations of the reachable positions of the mobile robot in different situations. (a) Obstacles on left and top; (b) obstacles on top left; (c) obstacles on top; (d) no obstacles.
Figure 5. Variations of the reachable positions of the mobile robot in different situations. (a) Obstacles on left and top; (b) obstacles on top left; (c) obstacles on top; (d) no obstacles.
Sensors 23 04647 g005
Figure 6. Flow chart of the biologically inspired complete coverage path planning algorithm based on Q-learning.
Figure 6. Flow chart of the biologically inspired complete coverage path planning algorithm based on Q-learning.
Sensors 23 04647 g006
Figure 7. The complete coverage path of the proposed algorithm in the 15 × 15 grid map.
Figure 7. The complete coverage path of the proposed algorithm in the 15 × 15 grid map.
Sensors 23 04647 g007
Figure 8. The complete coverage path of the boustrophedon algorithm in the 15 × 15 grid map.
Figure 8. The complete coverage path of the boustrophedon algorithm in the 15 × 15 grid map.
Sensors 23 04647 g008
Figure 9. The complete coverage path of the secondary area division algorithm in the 15 × 15 grid map.
Figure 9. The complete coverage path of the secondary area division algorithm in the 15 × 15 grid map.
Sensors 23 04647 g009
Figure 10. The complete coverage path of the proposed algorithm in the 30 × 20 grid map.
Figure 10. The complete coverage path of the proposed algorithm in the 30 × 20 grid map.
Sensors 23 04647 g010
Figure 11. The complete coverage path of the boustrophedon algorithm in the 30 × 20 grid map.
Figure 11. The complete coverage path of the boustrophedon algorithm in the 30 × 20 grid map.
Sensors 23 04647 g011
Figure 12. The complete coverage path of the secondary area division algorithm in the 30 × 20 grid map.
Figure 12. The complete coverage path of the secondary area division algorithm in the 30 × 20 grid map.
Sensors 23 04647 g012
Table 1. Comparison of complete coverage path data in the 15 × 15 grid map.
Table 1. Comparison of complete coverage path data in the 15 × 15 grid map.
AlgorithmCoverage Ratio (%)Repetition Ratio (%)Sheering Times
Improved algorithm in this paper1000.0035
Boustrophedon coverage algorithm1005.5232
Secondary region partitioning algorithm1002.7634
Table 2. Comparison of complete coverage path data in the 30 × 20 grid map.
Table 2. Comparison of complete coverage path data in the 30 × 20 grid map.
AlgorithmCoverage Ratio (%)Repetition Ratio (%)Sheering Times
Improved algorithm in this paper1000.0083
Boustrophedon coverage algorithm1006.2171
Secondary region partitioning algorithm1005.2572
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, X.; Han, L.; Gong, H.; Wu, Q. Biologically Inspired Complete Coverage Path Planning Algorithm Based on Q-Learning. Sensors 2023, 23, 4647. https://doi.org/10.3390/s23104647

AMA Style

Tan X, Han L, Gong H, Wu Q. Biologically Inspired Complete Coverage Path Planning Algorithm Based on Q-Learning. Sensors. 2023; 23(10):4647. https://doi.org/10.3390/s23104647

Chicago/Turabian Style

Tan, Xiangquan, Linhui Han, Hao Gong, and Qingwen Wu. 2023. "Biologically Inspired Complete Coverage Path Planning Algorithm Based on Q-Learning" Sensors 23, no. 10: 4647. https://doi.org/10.3390/s23104647

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop