Next Article in Journal
Genomic Approaches for Monogenic Kidney Diseases: A Comparative Review of Diagnostic Methods and Precision Medicine Implications
Previous Article in Journal
Isotropic ΙoT-Based Magnetic Flux Density Meter Implementation for ELF Field Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Path Planning for Automatic Berthing Using Ship-Maneuvering Simulation-Based Deep Reinforcement Learning

1
Department of Smart Environmental Energy Engineering, Changwon National University, Changwon 51140, Republic of Korea
2
Department of Naval Architecture and Marine Engineering, Changwon National University, Changwon 51140, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(23), 12731; https://doi.org/10.3390/app132312731
Submission received: 30 October 2023 / Revised: 23 November 2023 / Accepted: 25 November 2023 / Published: 27 November 2023
(This article belongs to the Topic Artificial Intelligence in Navigation)

Abstract

:
Despite receiving much attention from researchers in the field of naval architecture and marine engineering since the early stages of modern shipbuilding, the berthing phase is still one of the biggest challenges in ship maneuvering due to the potential risks involved. Many algorithms have been proposed to solve this problem. This paper proposes a new approach with a path-planning algorithm for automatic berthing tasks using deep reinforcement learning (RL) based on a maneuvering simulation. Unlike the conventional path-planning algorithm using the control theory or an advanced algorithm using deep learning, a state-of-the-art path-planning algorithm based on reinforcement learning automatically learns, explores, and optimizes the path for berthing performance through trial and error. The results of performing the twin delayed deep deterministic policy gradient (TD3) combined with the maneuvering simulation show that the approach can be used to propose a feasible and safe path for high-performing automatic berthing tasks.

1. Introduction

Since the early stages of modern shipbuilding, much attention has been paid to automated methods of ship navigation, particularly with the continuous advancements in artificial intelligence (AI). As a result, the number of autonomous ships has rapidly grown. Autonomous ship navigation offers substantial advantages in terms of safety, efficiency, reliability, and environmental sustainability. By harnessing advanced technologies such as sensor systems, data analysis, and artificial intelligence and by reducing the risk of human error, autonomous navigation systems ensure the safety of ship operations. These systems operate consistently and reliably, unhindered by human limitations, resulting in more predictable performance and fewer accidents. Autonomous ships can process vast amounts of data, enabling informed decision making, collision avoidance, and adaptation to changing conditions [1]. However, automatic ship berthing remains an extremely complex task, particularly under low-speed conditions where the hydrodynamic forces acting on the ship are highly nonlinear [2]. Controlling the ship becomes challenging and necessitates the expertise of an experienced commander. Numerous researchers have conducted extensive studies on the principles and algorithms for automatic ship berthing. Researchers have developed various ship control algorithms based on control theories and maneuverability assumptions [3,4,5,6,7,8,9]. These approaches proved effective under defined berthing conditions before the existence of AI. The development of AI algorithms has propelled the creation of algorithms and methods that enhance ship control performance, improve safety, and greatly reduce accidents in the marine industry. Many supervised learning algorithms based on neural networks have exhibited promising results with a high success rate, as in [10,11,12,13]. Applying AI algorithms eliminates the need to clearly understand the mathematical model of ships. However, acquiring a substantial number of labeled training data can be time-consuming and costly.
Unlike the aforementioned methods where the training dataset is not necessary, reinforcement learning techniques, which constitute an area of machine learning, allow the ship to learn and optimize its berthing maneuvers through interactions with a simulated environment. The application of RL in the automatic berthing task has shown good results when the ship can automatically learn the strategy and optimize the control policy to move to the berthing point [14,15].
In this paper, initial development of a novel path-planning algorithm for autonomous ship berthing that uses the latest technique in reinforcement learning, called twin delayed deep deterministic policy gradient (TD3), is proposed. The TD3 algorithm was introduced by Scott Fujimoto (2018) and is specifically designed for continuous action spaces. TD3 is an extension of deep deterministic policy gradients (DDPGs) and aims to address certain challenges and improve the stability of learning in complex environments. TD3 exploration algorithms allow the agent to explore the environment and gain new experiences that optimize rewards through trial and error. It employs two distinct value function estimators, which mitigate the overestimation bias and stabilize the learning process. Leveraging these two critics, TD3 provides more accurate value estimates and facilitates better policy updates. High performance and stability compared to other algorithms in the field of RL were shown in [16]. In combination with the MMG model, the solutions for ship-maneuvering motion simulation that were proposed by a research group maneuvering modeling group (MMG) in 1977 [17] suggest a feasible path, resulting in faster convergence and improved accuracy.
This article is organized into five parts. The first section introduces previous studies conducted in this field. Section 2 presents the equation of motion for the ship based on the MMG model along with hydrodynamic and interaction coefficients. Section 3 outlines the path-planning algorithm based on deep reinforcement learning TD3. Section 4 showcases and discusses simulation results for two berthing cases. Finally, Section 5 concludes this research.

2. Mathematical Model

2.1. Coordinated System

This paper focuses on the motions of USVs in horizontal planes only. Thus, two coordinate systems were defined for a maneuvering ship based on the right-hand rule, as shown in Figure 1. The earth-fixed coordinate system is O x y , where the origin is located on the water surface and the body-fixed coordinate system is o x b y b , where the origin is in the midship. The x and y axes point toward the ship’s bow and starboard, respectively. The heading angle ψ represents the angle between the x   and   x b axes.

2.2. Mathematical Model of USV

The motion equation with three degrees of freedom (3-DOF) is established based on Newton’s second law. In this paper, the berthing task assumed performing in the calm water, the disturbances from environments such as waves and wind in the port area are ignored due to the simplicity of the equation of motion. Thus, the heave, roll, and pitch motion are relatively small and not significantly affected in the equation of motion; thus, it can be neglected from the equation of motion. Furthermore, the low-speed condition made the 3-DOF motions sufficient to simulate the motion of the vehicle. Additionally, the main purpose of this paper is to focus on the path planning algorithm to generate the feasible path for the berthing task based on the maneuvering simulation. The application of the 3-DOF motion equations causes simplicity but retains characteristics of the system. The MMG model for the 3-DOF equation of motion suggested in [17] that the total ship force and moment can be divided into sub-components: the hull, thruster, and steering system. Thus, the 3-DOF motion equation is expressed as
m u ˙ v r x b G r 2 = X H + X P + X R m v ˙ + u r + x b G r ˙ = Y H + Y P + Y R I z z r ˙ + m x G v ˙ + u r = N H + N P + N R
where m is the mass of the ship; x b G is the longitudinal position of the center of gravity of the ship; u ,   v ,   r denote the surge, sway, and yaw velocities, respectively; the symbol “ ˙ ” in the head of the variable denotes the derivative of the variable with respect to time; I z z is the moment representing the mass moment of inertia with respect to the z-axis; X , Y , and   N represent the surge force, lateral force, and yaw moments, respectively, at the midship; and the subscripts H , P , and   R denote the hull, rudder, and propeller, respectively.
Due to the operation conditions in the berthing phase, the hydrodynamic forces and moments around the midship acting on the ship hull were investigated at a low speed with a wide range of drift angles [2]. Thus, the equation of motion considers some of the high-order hydrodynamic coefficients and the hydrodynamic forces and moments caused by the hull, which are expressed as follows:
X H = X u ˙ u ˙ + X u | u | u | u | + X v v v 2 + X r r r 2 + X v r v r Y H = Y v ˙ v ˙ + Y r ˙ r ˙ + Y v v + Y v v v v 3 + Y v v v v v v 5 + Y r r + Y r | r | r | r | + Y v v r v 2 r + Y v r r v r 2 N H = N v ˙ v ˙ + N r ˙ r ˙ + N v v + N u v u v + N v v v v 3 + N u v v v u v 3 + N r r + N r | r | r | r | + N v v r v 2 r + N v r r v r 2
The hydrodynamic coefficients were described under the Taylor series expansion in terms of surge velocity, sway velocity, and yaw rate.
The ship model was equipped with twin-propeller and twin-rudder systems [18]. The thruster model is expressed as follows:
X P = 1 t ρ D 4 n P 2 K T P J P P + n S 2 K T S J P S Y P = 0 N P = y P 1 t ρ D P 4 n P 2 K T P J P P n S 2 K T S J P S
where D is the propeller diameter; n is the propeller revolutions per minute; t denotes the thrust deduction factor; y P is the lateral position of the propeller from the centerline; and the superscripts S   and   P denote the side of the propeller (port and starboard).
The thrust coefficient K T is described as the function of the advanced ratio coefficient J P that was obtained through the propeller open water test
K T = k 0 + k 1 J P + k 2 J P 2
The parameter required for the estimation of thrust is given as
J P P , S = u P P , S n P , S D P u P P , S = 1 w P P , S u w P P , S = w P 0 exp C P v P 2 v P = v + x P r C P P = C P   and   C P S = C P +   when   β P > 0 C P P = C P +   and   C P S = C P   when   β P < 0
where the wake fraction at the propeller position w P was estimated using the wake fraction at the propeller in the straight motion w P 0 ; the geometrical inflow angle to the propeller position is denoted by β P ; C P +   and   C P describe the wake-changing coefficients for plus and minus β P due to lateral motion; and x P denotes the longitudinal position from the midship.
Forces and moments due to the steering system for the twin rudder were calculated based on the normal force F N and are expressed as
X R = 1 t R F N P + F N S sin δ Y R = 1 + a H F N P + F N S cos δ N R = x R + a H x H F N P + F N S cos δ y R 1 t R F N P F N S sin δ
where the normal force acting on the rudder is described as follows (Equation (7)).
F N P , S = 1 2 ρ A R U R P , S 2 f a sin α R P , S
The parameters required for estimating the rudder forces and moment during the maneuver are given as
U R P , S = u R P , S 2 + v R P , S 2 f a = 6.13 Λ Λ + 2.25 α R P , S = δ tan 1 v R P , S u R P , S u R P , S = ε u R P , S η 1 + κ 1 + 8 K T P , S π J P P , S 2 1 2 + 1 η v R P , S = γ R P , S v + l R r γ R P = γ R   and   γ R S = γ R +   when   β R > 0 γ R P = γ R +   and   γ R S = γ R   when   β R < 0
where F N is the normal rudder force; t R , a H , and x H are the steering resistance deduction factor, the rudder increase factor, and the position of an additional lateral force component, respectively; U R is the resultant rudder inflow velocity; f a is the rudder lift gradient coefficient; Λ is the rudder aspect ratio; α R is the effective inflow angle to the rudder; u R and v R are longitudinal and lateral inflow velocity components to the rudder; ε is a ratio of a wake fraction at the propeller and rudder position; γ is the flow straightening coefficient; and β R is the effective inflow angle to the rudder in the maneuvering motion.

2.3. Hydrodynamic and Interaction Coefficients

Previous studies were carried out by research groups at Changwon National University on hydrodynamic properties under operating conditions [19]. Experiments were conducted on the Korean autonomous surface ship (KASS) model, a ship model used in the project carried out by many universities and research institutes to develop autonomous ships. The main characteristics and the shape of the ship’s model are shown in Table 1 and Figure 2 The cross-comparison of the results between the previous studies and [20] shows similarities in the results. Figure 3 shows the comparison results of the turning maneuverability at a rudder angle of 35 degrees at three and six knots. Similarly, in order to obtain the feasible path for the automatic berthing task, the high-accuracy equation of motion and hydrodynamic coefficients of the USV should be investigated carefully using the Korean autonomous surface ship (KASS) model. In this paper, the hydrodynamic coefficients were estimated using the captive model test at Changwon National University and compared with the CFD method as presented in [21]. The coefficients relative to only the surge velocity were estimated through the resistance test. The hydrodynamic coefficients related to surge and sway velocity for forces and moments were estimated through a static drift test. The hydrodynamic coefficients related to the yaw rate were estimated using the circular motion test and the hydrodynamic coefficients related to the combined effect of sway velocity and yaw rate were estimated using the combined circular motion with drift test. The added mass and interaction coefficients were selected from [21]. The summary of hydrodynamics and interaction coefficients are shown in Table 2 and Table 3.

2.4. Maneuverability

To define the problem, it is necessary to assess the maneuverability of the ship. Understanding the maneuverability makes it reasonable to determine where to start the automatic berthing process. The maneuvering simulation under low speed (1 knot) was conducted to investigate the ship-maneuvering characteristics in the port environment. Figure 4 shows the trajectory of the turning circle test at 35 degrees of a rudder angle at 1 knot. The simulation results show that this can easily turn into a range of approximately 3 L P P . Thus, the berthing area should be greater than three times that of L P P from the berthing point. The maneuvering characteristics are shown in Table 4.

3. Path-Planning Approach

In the last few decades, significant advancements have been made in the field of artificial intelligence, particularly reinforcement learning, a subfield of machine learning, which falls under the machine learning category. Reinforcement learning involves training data by assigning rewards and punishments based on behavior and state. Unlike supervised and semi-supervised learning, reinforcement learning does not rely on pairs of input data or true results and it does not explicitly evaluate near-optimal actions as true or false. As a result, reinforcement learning offers a solution to tackle complex problems, including the control of robots, self-driving cars, and even applications in the aerospace industry. A noteworthy advancement in reinforcement learning is the introduction of a twin delayed deep deterministic policy gradient in 2018. TD3 is an effective model-free policy reinforcement learning method. The TD3 agent is an actor–critic reinforcement learning agent that optimizes the expected long-term reward. Specifically, TD3 builds upon the success of the deep deterministic policy gradient algorithm developed in 2016. DDPG remains highly regarded and successful in the continuous action space, finding extensive applications in fields such as robotics and self-driving systems.
However, like many algorithms, DDPG has its limitations, including instability and the need for fine-tuning hyperparameters for each task. Estimation errors gradually accumulate during training, leading to suboptimal local states, overestimation, or severe forgetfulness on the part of the agent. To address these issues, TD3 was developed with a focus on reducing the overestimation bias prevalent in previous reinforcement learning algorithms. This is achieved through the incorporation of three key features:
  • The utilization of twin critic networks, which work in pairs.
  • Delayed updates of the actor.
  • Action noise regularization.
By implementing these features, TD3 aims to enhance the stability and performance of reinforcement learning algorithms, ultimately improving their applicability in various domains.

3.1. Conception

The path-planning algorithm in this paper was performed using the KASS model mentioned in Section 2.3. The port selected was the Busan port, whose geometry is shown in Figure 5. The objective of this result is to use the TD3 (the pseudocode as shown in Algorithm 1) for training the model that can generate the path for the berthing process. First, the TD3 algorithm trains the model with the combination of maneuvering simulations suggested by the MMG model. This approach allows for the integration of realistic ship motion dynamics into the training process. Then, this model is used to generate the desired path for the berthing task by inputting the state of the ship and predicting the control signal n (propeller speed) and δ (rudder angle), based on the input ship state s ( x , y , ψ , u , v ,   and   r ) . The concept of path planning for automatic berthing tasks is shown in Figure 6.
Algorithm 1: Pseudocode TD3 Algorithm
1Initialize the critic network Q ϕ 1 , Q ϕ 2 and actor-network μ θ with random parameter ϕ 1 , ϕ 2 , θ
2Initialize the target parameter to the main parameter
θ t arg θ , ϕ t arg 1 ϕ 1 , ϕ t arg 2 ϕ 2
3For t = 0 to T 1 do:
4    Observe the state s of the environment and choose the action
a = c l i p μ θ s + ε , a l o w , a h i g h   where   ε ~ N
5    Execute action a in the TD3 environment to observe the new state s , reward r , and done signal that gives the signal to stop training for this step.
6    Store training set s , a , r , s , d in replay buffer D
7    If s taking to the goal point, reset the environment state
8    If it is time for an update:
For j in range (custom decided) do:
9       Randomly sample a batch of transitions
B = s , a , r , s , d   from   D
10      Compute Target actions: a s = c l i p μ θ t arg s + c l i p ( ε , c , c ) , a l o w , a h i g h   where   ε ~ N , σ
11        Compute targets:
y r , s , d = r + γ 1 d min i = 1 , 2 Q ϕ t arg i s , a s
12       Update Q-functions using gradient descent:
            ϕ i 1 B s , a , r , s , d B Q ϕ i s , a y r , s , d 2                 where   i = 1 , 2
13        If j mod policy delay == 0 then:
Update policy by the one-step deterministic policy gradient ascent using
θ 1 B s B Q ϕ 1 s , μ θ s
14                Update target networks:
ϕ t arg . i ρ ϕ t arg . i + 1 ρ ϕ i   where   i = 1 , 2 θ t arg ρ θ t arg + 1 ρ θ
15       End if
End for
End if
End until convergence

3.2. Setting for Reinforcement Learning

In this section, the parameters and variables for the TD3 algorithm to deal with the automatic berthing task were set as follows:
  • Observation space and state: The observation space and state were defined as the set of physical velocity, position, and orientation. The state vector s ( x , y , ψ , u , v ,   and   r ) includes the position x , y as the element. The orientation ψ is the heading angle. The linear velocity is u , v and the angular velocity is r ;
  • Action: The control action includes the control input of the thruster (revolution of propeller) and steering (rudder angle) system. The action signal is continuous in the range [−1, 1], where [−1, 1] = [−300, 100] rpm represents the thrust system and [−1, 1] where [−1, 1] = [−35, 35] is the degrees for the steering system.
  • Reward function: This plays a crucial role in the design of a reinforcement learning application. It serves as a guide for the network training process and helps optimize the model’s performance throughout each episode. If the reward function does not accurately capture the objectives of the target task, the model may struggle to achieve desirable performance.
    r S = r i r i 1 r = r D i s t + r L i n V e l + r A n g V e l + r H e a d i n g r D i s t = 100 x 2 + y 2 r H e a d i n g = 1000 ψ T arg e t ψ r L i n V e l = 2000 u 2 + v 2 r A n g V e l = 1000 r
In this paper, based on the boundary state, the weights 100, 1000, 2000, and 1000 were assigned to the weight of distance, heading, linear, and angular velocity, respectively. The reward value was described as the sum of the reward in each time step. It received a positive value if the state variable changed to the required value and vice versa. Furthermore, for the faster convergence of rewards in the first stage of the berthing task, the distance is more prior to the target state than the speed and heading. Thus, the reward function of distance is multiplied by the distance coefficients. It is (1.1-Distance coefficients) in the case of the reward function of the heading, resultant velocity, and yaw rate. The reward coefficients are described in Figure 7.
  • Environment: The environment receives the input as the control input and the state then returns the ship’s new state and the reward for this action. The environment function was built based on a maneuvering simulation that uses the MMG model as a mathematical simulation.
  • Agent: The hyperparameters for the TD3 model were selected as follows: the number of hidden layers was set as two layers with 512 units for each. The learning rate for the actor and critic networks α and β was set to 0.0001. The discount factor γ was 0.99. The soft update coefficient τ was 0.005. The batch size was 128. This training process was set to 20,000/50,000 steps for the warmup of the model with the exploration noise set in Table 5.

3.3. Boundary Conditions

The simulations were performed at Busan port, with the satellite capture as in Figure 7. The geometry of this port was simplified as Figure 8. Considering the geometry of this port, two cases of berthing were selected to investigate the path-plan-generating system.
Case 1: Parallel berthing task in a 30.5 × 202 m water area. In this case, the ship’s initial states were generated randomly, as described in Table 6. The berthing point was assumed, as shown in Table 7. The berthing task can be considered a success if the ship state is in the range of the values described in Table 7. The boundary area is described as a Figure 8a and Table 8.
Case 2: Perpendicular berthing task in a 42 × 170.5 m water area. In this case, the ship’s initial states were generated randomly, as described in Table 9. The berthing point was assumed, as shown in Table 10. The berthing task can be considered a success if the ship state is in the range of the values described in Table 10. The boundary area is described as a Figure 8b and Table 11.

4. Simulation Results and Discussion

The USV is considered to have successful berthing if it approaches the target berthing point with an error within the allowable range established in Section 3.3. The training process stops when the number of training episodes reaches 50,000.
Figure 9 and Figure 10 sequentially present a set of information that includes the trajectory, surge, sway, yaw rate, heading angle, and control input as the propeller revolution and rudder angle. The USV starts from the random state shown in Section 3.3. Under the automatic control of the TD3 model, the ship has successfully berthed at the target location. The ship states gradually change from the initial state to the required range. In particular, the combination with the adjusted reward function made the ship berthing process happen faster and more optimally. The time series of surge velocity was shown in the first phase of the berthing process. The ship’s speed was increased to reduce the distance to the berthing point. However, the sway velocity and yaw rate did not change much in this phase. In the last phase of the berthing process, the importance of surge velocity, yaw rate, and heading angle is much more important than the distance in the first phase. So, the change in this value seems more sudden to adapt to the required value of the berthing process.
Figure 11a,b shows the learning performance of TD3 case 1 with 50,000 episodes. Although the warmup episode number is 20,000, the average reward shows that the model seems to be successfully berthing and stable before the warmup phase is finished. This demonstrates that TD3 provides good reinforcement learning for the automatic berthing process. To account for the difference in penalty values affecting the training process, too high a penalty will cause the model to misjudge the states, for instance, in cases where the vessel has moved reasonably close to the berthing position. However, it then collides with the wall and receives a heavy point deduction, causing the model to judge that process as wrong and try actions other than that process. This makes the training process longer. These results show that the method proposed in this paper has a high performance and success rate with a low penalty value.
The simulation results show that the combination of TD3 and the maneuvering simulation proposes a powerful and accurate system for the automatic berthing process. Comparing the shape of the average reward to the results shown in [15] demonstrates that the stability of the TD3 algorithm is better than the older algorithm in the field of reinforcement learning. In particular, the method proposed in this paper is easier to apply because of the ability to learn, explore, and optimize the policy automatically. It can be used for another ship model if we know the ship’s hydrodynamic characteristics.
However, the limitations in this paper are evident. Firstly, the simplification of the port condition: due to the simplicity of the model, the effect of the disturbance was ignored in the simulation. This causes non-accuracy if there is wind or waves. The second limitation of this approach is the simplification of the obstacles and port geometry. It has a significant effect on the determination of the initial state and the berthing point. The presence of moving obstacles can make the training time increase significantly. So, the method proposed in this paper should only be used in the determined port and not for moving obstacles. Finally, this paper proposed the initial development of the path planning system. The results are performed based on maneuver simulation. The accuracy and performance of this approach in real-world operations need to be carefully considered and evaluated.
Although the approach in this paper uses the newest technique in reinforcement learning at present, the performance of this method should be investigated and compared carefully.

5. Conclusions and Remarks

In this study, the Korea autonomous surface ship (KASS) model was selected as the target ship to perform the training for path planning for the autonomous berthing task. A mathematical model and the hydrodynamics coefficients suggested in previous research conducted at Changwon National University provided an accurate model for solving the motion of a slow ship. By performing the path-planning algorithm based on the combination of TD3 and a maneuvering simulation, the automatic berthing task could be conducted with the automatic berthing problem and stable performance using reinforcement learning.
Even though the high performance of the path-planning system was shown, the complex environmental disturbance in the port area needs to be included in the model. It takes more time to train the model but is a necessary factor in the real situation. Additionally, several algorithms based on control theory must be considered for faster convergence.

Author Contributions

Conceptualization, H.K.Y. and A.K.V.; methodology, H.K.Y., A.K.V. and T.L.M.; software, A.K.V.; validation, A.K.V. and T.L.M.; formal analysis, A.K.V. and T.L.M.; investigation, A.K.V. and T.L.M.; resources, T.L.M.; data curation, A.K.V. and T.L.M.; writing—original draft preparation, A.K.V.; writing—review and editing, H.K.Y. and T.L.M.; visualization, A.K.V.; supervision, H.K.Y.; project administration, H.K.Y.; funding acquisition, H.K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Development of Autonomous Ship Technology (PJT201313, Development of Autonomous Navigation System with Intelligent Route Planning Function), funded by the Ministry of Oceans and Fisheries (MOF, Korea).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article (Tables and Figures).

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

ItemUnit
α R -Effective inflow angle to the rudder
β r a d Drift angle of the ship
β P r a d Inflow angle to the propeller
β R r a d Inflow angle to the rudder
γ R -Flow straightening coefficient of the rudder
δ r a d Rudder angle
η -Propeller diameter to the rudder span ratio
Λ -Rudder aspect ratio
κ -The experimental coefficient for longitudinal inflow velocity to the rudder
m 3 Displacement of the ship
ψ r a d Heading angle of the ship
ρ k g / m 3 Water density
ε -Ratio of the wake fraction at the propeller to the rudder
A R m 2 Rudder profile area
a H -Increase factor of the rudder force
B m Breadth of ship
B R m Average rudder chord length
C P -Experimental constant due to wake characteristic in maneuvering
D P m Diameter of propeller
d m Draft of ship
F N N Normal force of the rudder
F X , F Y N Surge and sway force acting on the ship
f α -Lift gradient coefficients of the rudder
H R m Span of rudder
I Z k g m 2 Moment of inertial of ship
J P -Advanced ratio of propeller
K T -Propeller thrust open water characteristic
k 0 , k 1 , k 2 -Coefficients relative to K T
L P P m Ship length between two perpendicular
l R m Effective longitudinal length of the rudder position
M Z N m Yaw moment acting on the ship
m k g Ship mass
n P , n S r p m Propeller revolutions per minute (rpm)
O x y z -Earth-fixed coordinate system
o x b y b z b -Body-fixed coordinate system
R 0 -Resistance of the ship in straight motion (-)
r d e g / s Yaw rate
r Reward
T N Longitudinal propeller force
t s Time
t P -Thrust deduction factor
t R -Steering deduction factor
U m / s Resultant velocity
U 0 m / s Initial resultant velocity
U R m / s Resultant inflow velocity to the rudder
u , v m / s Longitudinal and lateral velocity of the ship in the body-fixed coordinate system
u R , v R m / s Longitudinal and lateral inflow velocity to rudder position
w P -Wake coefficient at the propeller in maneuvering motion
w P 0 -Wake coefficient at the propeller at straight motion
w R -Wake coefficient at the rudder position
X ,   Y ,   N N ,   N ,   N m Surge, sway force, and yaw moment around the midship
X H ,   Y H ,   N H N ,   N ,   N m Surge, sway force, and yaw moment acting on ship’s Hull
X P ,   Y P ,   N P N ,   N ,   N m Surge, sway force, and yaw moment due to the propeller
X R ,   Y R ,   N R N ,   N ,   N m Surge, sway force, and yaw moment due to the rudder
x G m Longitudinal position of the center of gravity
x H m Longitudinal position of the acting point of addition lateral force
x P m Longitudinal position of the propeller
x R m Longitudinal position of the rudder

Abbreviations

Item
AIArtificial Intelligence
CFDComputational Fluid Dynamics
DDPGDeep Deterministic Policy Gradients
DRLDeep Reinforcement Learning
MMGManeuvering Modeling Group
RLReinforcement Learning
TD3Twin Delayed DDPG (a variant of the DDPG algorithm)
USVUnmanned Surface Vehicle

References

  1. Chaal, M.; Ren, X.; BahooToroody, A.; Basnet, S.; Bolbot, V.; Banda, O.A.V.; van Gelder, P. Research on risk, safety, and reliability of autonomous ships: A bibliometric review. In Safety Science (Vol. 167); Elsevier B.V.: Amsterdam, The Netherlands, 2023. [Google Scholar] [CrossRef]
  2. Oh, K.G.; Hasegawa, K. Low speed ship manoeuvrability: Mathematical model and its simulation. In Proceedings of the International Conference on Offshore Mechanics and Arctic Engineering—OMAE, Nantes, France, 9–14 June 2013; p. 9. [Google Scholar] [CrossRef]
  3. Shouji, K. An Automatic Berthing Study by Optimal Control Techniques. IFAC Proc. Vol. 1992, 25, 185–194. [Google Scholar] [CrossRef]
  4. Skjåstad, K.G.; Barisic, M. Automated Berthing (Parking) of Autonomous Ships. Ph.D. Thesis, NTNU, Trondheim, Norway, 2018. [Google Scholar]
  5. Mizuno, N.; Uchida, Y.; Okazaki, T. Quasi real-time optimal control scheme for automatic berthing. IFAC-Pap. 2015, 28, 305–312. [Google Scholar] [CrossRef]
  6. Nguyen, V.S.; Im, N.K. Automatic ship berthing based on fuzzy logic. Int. J. Fuzzy Log. Intell. Syst. 2019, 19, 163–171. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Zhang, M.; Zhang, Q. Auto-berthing control of marine surface vehicle based on concise backstepping. IEEE Access 2020, 8, 197059–197067. [Google Scholar] [CrossRef]
  8. Sawada, R.; Hirata, K.; Kitagawa, Y.; Saito, E.; Ueno, M.; Tanizawa, K.; Fukuto, J. Path following algorithm application to automatic berthing control. J. Mar. Sci. Technol. 2021, 26, 541–554. [Google Scholar] [CrossRef]
  9. Wu, G.; Zhao, M.; Cong, Y.; Hu, Z.; Li, G. Algorithm of berthing and maneuvering for catamaran unmanned surface vehicle based on ship maneuverability. J. Mar. Sci. Eng. 2021, 9, 289. [Google Scholar] [CrossRef]
  10. Im, N.; Seong Keon, L.; Hyung Do, B. An Application of ANN to Automatic Ship Berthing Using Selective Controller. Int. J. Mar. Navig. Saf. Sea Transp. 2007, 1, 101–105. [Google Scholar]
  11. Ahmed, Y.A.; Hasegawa, K. Automatic ship berthing using artificial neural network trained by consistent teaching data using nonlinear programming method. Eng. Appl. Artif. Intell. 2013, 26, 2287–2304. [Google Scholar] [CrossRef]
  12. Im, N.; Hasegawa, K. Automatic ship berthing using parallel neural controller. IFAC Proc. Vol. 2001, 34, 51–57. [Google Scholar] [CrossRef]
  13. Im, N.K.; Nguyen, V.S. Artificial neural network controller for automatic ship berthing using head-up coordinate system. Int. J. Nav. Archit. Ocean. Eng. 2018, 10, 235–249. [Google Scholar] [CrossRef]
  14. Marcelo, J.; Figureueiredo, P.; Pereira, R.; Rejaili, A. Deep Reinforcement Learning Algorithms for Ship Navigation in Restricted Waters. Mecatrone 2018, 3, 151953. [Google Scholar] [CrossRef]
  15. Lee, D. Reinforcement Learning-Based Automatic Berthing System. arXiv 2021, arXiv:2112.01879. [Google Scholar]
  16. Fujimoto, S.; van Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. Int. Conf. Mach. Learn. 2018, 80, 1587–1596. [Google Scholar]
  17. Yasukawa, H.; Yoshimura, Y. Introduction of MMG Standard Method for Ship Maneuvering Predictions. J. Mar. Sci. Technol. 2015, 20, 37–52. [Google Scholar] [CrossRef]
  18. Khanfir, S.; Hasegawa, K.; Nagarajan, V.; Shouji, K.; Lee, S.K. Manoeuvring characteristics of twin-rudder systems: Rudder-hull interaction effect on the manoeuvrability of twin-rudder ships. J. Mar. Sci. Technol. 2011, 16, 472–490. [Google Scholar] [CrossRef]
  19. Vo, A.K.; Mai, T.L.; Jeon, M.; Yoon, H.k. Experimental Investigation of the Hydrodynamic Characteristics of a Ship due to Bank Effect. Port. Res. 2022, 46, 294–301. [Google Scholar] [CrossRef]
  20. Kim, D.J.; Choi, H.; Kim, Y.G.; Yeo, D.J. Mathematical Model for Harbour Manoeuvres of Korea Autonomous Surface Ship (KASS) Based on Captive Model Tests. In Proceedings of the Conference of Korean Association of Ocean Science and Technology Societies, Incheon, Republic of Korea, 13–14 May 2021. [Google Scholar]
  21. Vo, A.K. Application of Deep Reinforcement Learning on Ship’s Autonomous Berthing Based on Maneuvering Simulation. Ph.D. Thesis, Changwon National University, Changwon, Republic of Korea, 2022. [Google Scholar]
Figure 1. Coordinate system of the twin-propeller and twin-rudder ship model.
Figure 1. Coordinate system of the twin-propeller and twin-rudder ship model.
Applsci 13 12731 g001
Figure 2. Geometry of an autonomous surface ship (KASS).
Figure 2. Geometry of an autonomous surface ship (KASS).
Applsci 13 12731 g002
Figure 3. Simulation results of turning trajectories at a rudder angle of 35 degrees at three and six knots [20].
Figure 3. Simulation results of turning trajectories at a rudder angle of 35 degrees at three and six knots [20].
Applsci 13 12731 g003
Figure 4. Simulation results of turning trajectories at a rudder angle of 35 degrees.
Figure 4. Simulation results of turning trajectories at a rudder angle of 35 degrees.
Applsci 13 12731 g004
Figure 5. Satellite image of the Busan port.
Figure 5. Satellite image of the Busan port.
Applsci 13 12731 g005
Figure 6. Concept of the path-planning algorithm.
Figure 6. Concept of the path-planning algorithm.
Applsci 13 12731 g006
Figure 7. Reward coefficients.
Figure 7. Reward coefficients.
Applsci 13 12731 g007
Figure 8. Geometry of the port and berthing situation: (a) parallel berthing task (case 1) and (b) perpendicular berthing task (case 2).
Figure 8. Geometry of the port and berthing situation: (a) parallel berthing task (case 1) and (b) perpendicular berthing task (case 2).
Applsci 13 12731 g008
Figure 9. Simulation results in a parallel berthing task (case 1).
Figure 9. Simulation results in a parallel berthing task (case 1).
Applsci 13 12731 g009
Figure 10. Simulation results in the perpendicular berthing task (case 2).
Figure 10. Simulation results in the perpendicular berthing task (case 2).
Applsci 13 12731 g010
Figure 11. Learning performance of TD3: (a) parallel berthing (case 1) and (b) perpendicular berthing task (case 2).
Figure 11. Learning performance of TD3: (a) parallel berthing (case 1) and (b) perpendicular berthing task (case 2).
Applsci 13 12731 g011
Table 1. Principal dimensions.
Table 1. Principal dimensions.
Item (Unit)Value
Length perpendicular, L p p ( m )22.000
Breadth, B ( m )6.000
Draft, T ( m )1.250
Displacement Volume, V ( m 3 )86.681
Rudder area, A R ( m 2 )0.518
Rudder span, H R ( m )0.900
Propeller diameter, D P ( m )0.950
Table 2. Hydrodynamic force and moment coefficients.
Table 2. Hydrodynamic force and moment coefficients.
Hull   ( × 1 0 5 )
X u ˙ −81 Y v ˙ −1034 Y v ˙ 64
X u u −627 Y r ˙ −126 Y r ˙ −33
X v v −407 Y v −2610 N v −130
X r r 675 Y v v v −3530 N u v −513
X v r 226 Y v v v v v 3080 N v v v −2
Y r 390 N u v v v −138
Y r r −47 N r −178
Y v r r −2170 N r r −253
Y v v r −3590 N v r r −420
N v v r −1830
Table 3. Interaction coefficients.
Table 3. Interaction coefficients.
Propeller and Rudder
1 t R 0.934 1 + a H 0.702 γ R + 0.342
ε 0.960 C P + −2.713 γ R 0.634
κ 0.695 C P 11.211
Table 4. Turning maneuverability characteristics.
Table 4. Turning maneuverability characteristics.
Starboard TurningPort Turning
Advance ( L P P )2.6812.680
Transfer ( L P P )1.1141.115
Turning Radius ( L P P )2.6250.623
Tactical diameter ( L P P )2.9352.932
Table 5. Exploration rates.
Table 5. Exploration rates.
ItemValue (Exploration Rate ϵ)
Step [0–5000]0.5
Step [5000–10,000]:0.4
Step [10,000–15,000]0.3
Step [15,000–20,000]0.2
Step [20,000–50,000]0.1
Table 6. Initial state values (case 1).
Table 6. Initial state values (case 1).
ItemValue
x 0   ( m ) [−20, 20]
y 0   ( m ) [−10, 10]
ψ 0   ( ° ) [−5, 5]
u 0   ( k n o t ) 1
v 0   ( k n o t ) 0
r 0   ( ° / s ) 0
Table 7. Target point (case 1).
Table 7. Target point (case 1).
ItemValue
x   ( m ) 180 ± 2
y   ( m ) 20 ± 0.5
ψ   ( ° ) 0 ± 3
u   ( m ) 0 ± 0.1
v   ( m ) 0 ± 0.05
r   ( ° / s ) 0 ± 1
Table 8. Boundary values (case 1).
Table 8. Boundary values (case 1).
ItemValue
x   ( m ) [−20, 182]
y   ( m ) [−10, 20.5]
Table 9. Initial state values (case 2).
Table 9. Initial state values (case 2).
ItemValue
x 0   ( m ) [−20, 20]
y 0   ( m ) [−10, 10]
ψ 0   ( ° ) [−5, 5]
u 0   ( k n o t ) 1
v 0   ( k n o t ) 0
r 0   ( ° / s ) 0
Table 10. Target point (case 2).
Table 10. Target point (case 2).
ItemValue
x   ( m ) 150 ± 0.5
y   ( m ) 30 ± 2
ψ   ( ° ) 90 ± 3
u   ( m / s ) 0 ± 0.1
v   ( m / s ) 0 ± 0.05
r   ( ° / s ) 0 ± 1
Table 11. Boundary values (case 2).
Table 11. Boundary values (case 2).
ItemValue
x   ( m ) [−20, 150.5]
y   ( m ) [−10, 32]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vo, A.K.; Mai, T.L.; Yoon, H.K. Path Planning for Automatic Berthing Using Ship-Maneuvering Simulation-Based Deep Reinforcement Learning. Appl. Sci. 2023, 13, 12731. https://doi.org/10.3390/app132312731

AMA Style

Vo AK, Mai TL, Yoon HK. Path Planning for Automatic Berthing Using Ship-Maneuvering Simulation-Based Deep Reinforcement Learning. Applied Sciences. 2023; 13(23):12731. https://doi.org/10.3390/app132312731

Chicago/Turabian Style

Vo, Anh Khoa, Thi Loan Mai, and Hyeon Kyu Yoon. 2023. "Path Planning for Automatic Berthing Using Ship-Maneuvering Simulation-Based Deep Reinforcement Learning" Applied Sciences 13, no. 23: 12731. https://doi.org/10.3390/app132312731

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop