Next Article in Journal
Multi-Objective Discrete Brainstorming Optimizer to Solve the Stochastic Multiple-Product Robotic Disassembly Line Balancing Problem Subject to Disassembly Failures
Previous Article in Journal
An Inventory Model with Advertisement- and Customer-Relationship-Management-Sensitive Demand for a Product’s Life Cycle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning-Based Lane Change Decision for CAVs in Mixed Traffic Flow under Low Visibility Conditions

1
Department of Traffic Information and Control Engineering, Jilin University, Changchun 130022, China
2
China Academy of Transportation Sciences, Beijing 100029, China
3
Jilin Engineering Research Center for Intelligent Transportation System, Changchun 130022, China
4
Qingdao Transportation Public Service Center, Qingdao Municipal Transport Bureau, Qingdao 266061, China
*
Authors to whom correspondence should be addressed.
Mathematics 2023, 11(6), 1556; https://doi.org/10.3390/math11061556
Submission received: 5 March 2023 / Revised: 19 March 2023 / Accepted: 20 March 2023 / Published: 22 March 2023
(This article belongs to the Section Fuzzy Sets, Systems and Decision Making)

Abstract

:
As an important stage in the development of autonomous driving, mixed traffic conditions, consisting of connected autonomous vehicles (CAVs) and human-driven vehicles (HDVs), have attracted more and more attention. In fact, the randomness of human-driven vehicles (HDV) is the largest challenge for connected autonomous vehicles (CAV) to make reasonable decisions, especially in lane change scenarios. In this paper, we propose the problem of lane change decisions for CAV in low visibility and mixed traffic conditions for the first time. First, we consider the randomness of HDV in this environment and construct a finite state machine (FSM) model. Then, this study develops a partially observed Markov decision process (POMDP) for describing the problem of lane change. In addition, we use the modified deep deterministic policy gradient (DDPG) to solve the problem and get the optimal lane change decision in this environment. The reward designing takes the comfort, safety and efficiency of the vehicle into account, and the introduction of transfer learning accelerates the adaptation of CAV to the randomness of HDV. Finally, numerical experiments are conducted. The results show that, compared with the original DDPG, the modified DDPG has a faster convergence velocity. The strategy learned by the modified DDPG can complete the lane change in most of the scenarios. The comparison between the modified DDPG and the rule-based decisions indicates that the modified DDPG has a stronger adaptability to this special environment and can grasp more lane change opportunities.

1. Introduction

In recent years, connected autonomous vehicle (CAV) has been widely discussed in the world. Sensing, decision-making, planning, control and other related autonomous driving technologies have also entered a period of rapid development. Decision technology is one of the most critical modules. It receives information from the perception module and gives guidance to planning and control modules. Therefore, decision-making technology is equivalent to the brain of CAV and plays a key role in the safe driving of the vehicle [1]. Lane change is the main scenario for CAV decision-making, including the determination of lane change intention and execution time. In this paper, we assume that the vehicle has a strong intent to change lanes. The CAV needs to choose the right time to change lanes and perform the lane change. During a lane change, the CAV needs to pay attention to the real-time state of the other vehicles in the area that are affected by the lane change, including velocity, acceleration and relative position. In particular, the success of the lane change is often directly determined by predicting the driving intention of vehicles behind the target lane.
The judgment of lane change timing is the key to a successful lane change. The basis of the judgment includes some hard rules [2], such as the calculation of real-time benefits [3] and the existing experience of the lane change [4,5].
In mixed traffic scenarios, drivers are stochastic, and relying on hard rules may cause the CAV to miss some lane change opportunities. Therefore, lane change decisions based on hard rules are conservative. Benefit calculation-based decision-making usually requires real-time interaction with the driver on the road; however, this approach is difficult to apply in low-visibility weather such as haze, fog, etc. There are two main reasons for this: on the one hand, the sensing sensors (LIDAR, cameras, etc.) may be affected to varying degrees [6], and on the other hand, the low-visibility environment can affect the driver’s view. For the lane change experience, it is difficult to collect sufficient marker data from skilled drivers, especially in low-visibility scenarios. Therefore, it is urgent to develop a lane change decision algorithm for autonomous driving under low visibility mixed traffic conditions.
The problem of lane change decisions for CAV under mixed traffic conditions has also been widely studied [7,8,9,10]. In the existing research, researchers have used hard rule-based lane change decision methods, such as FSM [11], decision trees (DT) [12] and custom rules [13]. Other researchers have used machine learning methods to guide lane change decisions by learning from successful lane change experiences. These methods are the support vector machines (SVM) [14], XGBoost [4], Gaussian mixture density hidden Markov model (GMM-HMM) [15] and neural networks [16]. There are also many researchers who focus on decision methods for computing benefits, especially game theoretic methods based on real-time interactions [17]. Reinforcement learning, which combines the benefits of computing and machine learning in lane change strategies, has also been extensively studied [18,19]. All the above studies address lane change decisions for CAV in mixed traffic conditions in normal weather. To the best of our knowledge, no studies have been conducted on the lane change decision of CAV in low visibility and mixed conditions.
In this study, we formulated the lane change decision for CAV as a partially observed problem and solved it using modified reinforcement learning methods. First, we considered the effects of trust and visibility on drivers and proposed an FSM to characterize drivers’ driving in low visibility and mixed traffic conditions. Second, we used a POMDP to represent the problem of changing lanes for CAV in low visibility and mixed traffic conditions. Third, we performed the solution with a modified reinforcement learning algorithm. Finally, we tested and verified the effectiveness of the modified algorithm.
The main contributions of this research are as follows: First, it introduces a new scenario for lane changing in CAV in mixed traffic and low visibility conditions. As far as we know, this is the first study to address the lane change decision problem in such conditions. Second, the study identifies four car following states that drivers experience and combines the weights of safety and space factors to represent a driver’s response to a vehicle ahead. These five states are then integrated into an FSM model. Third, the researchers use POMDP to formulate the entire lane change decision problem for CAV and develop a modified DDPG algorithm to solve this problem.
The remainder of this paper is organized as follows: Section 2 reviews related research on the lane change decision-making in mixed traffic conditions for CAV. Section 3 presents the driver model formulation and problem description. Section 4 describes our proposed solution methodology, the modified DDPG. In Section 5, numerical experiments are conducted to illustrate the effectiveness of our algorithm. This paper is concluded in Section 6 with the summary and possible future research.

2. Literature Review

In order to draw a clear distinction between this study and previous research, the literature on rule-based, game-theoretic, machine-learning and reinforcement-learning is reviewed. The influence of low visibility on the perception module of CAV and the behavior characteristics of drivers under low visibility and mixed traffic conditions are also introduced in this section.
The existing research on rule-based lane change decisions is adequate. In these studies, lane changes are guided by a set of deterministic rules [20]. The main purpose is to ensure the absolute safety of lane change by judging whether vehicles meet lane change conditions. It is based on the idea of absolute safety that Mobileye proposes a liability-sensitive safety policy [21]. The earliest rule-based lane change decision models were mainly concerned with the acceptance of lane change gaps [22,23]. The Gipps model [22] defines the lane change process as a series of rules that use a decision tree and outputs the result as a binary classification result. Of course, the model is no longer satisfied with the complex vehicle composition and complex human-vehicle interaction behavior in the traffic environment. Another typical representative of the rule-based lane change model is the MOBIL model [24], which determines whether to change lanes by setting utility incentives and safety rules that apply to various rules of following and free lane change and forced lane change behavior. Of course, there is also a class of minimum safe distance models model, which determines the safe lane change condition by specifying the minimum safe distance rule between vehicles. At present, the biggest shortcoming of this type of rule-based model is that the driving environment conditions are relatively simple, and the influence of the driver’s driving behavior on lane changing is not much considered. In fact, these rules are the basic elements for making lane change decisions, and many new decision methods are further researched based on these rules.
Although these rules guarantee absolute safety for lane changes, they require a large enough clearance, which is unrealistic in the case of traffic congestion or accidents. Therefore, on the basis of accepting gaps, some studies have extended lane change rules, taking into account the interaction between vehicles and the variability of individual drivers [25,26,27,28,29,30]. Game theory is commonly used to reflect driver differences in inter-vehicle interactions [31]. However, in low visibility, even if the sensing module of the CAV works properly, the driver’s restricted field of view causes him/her not to interact properly with the surrounding vehicles, therefore, the result between vehicles cannot be trusted.
With the development of artificial intelligence, many scholars try to improve the safety of lane change decisions by machine learning methods. Chen et al. [32] used a random forest classifier to extract key features from vehicle trajectory data and then used fault analysis and k-mean clustering algorithms to determine the risk level of vehicle lane-changing behavior. Xu et al. [5] developed a convolutional neural network (CNN) model based on the lane-changing image dataset to estimate the driver’s intention to change lanes. Jin et al. [16] proposed a Gauss mixture hidden Markov model (GM-HMM) lane change decision model, which is experimentally shown to have a high similarity rate to the actual lane change behavior. Gindele et al. [33] constructed a filter to estimate the behavior of traffic participants and predict their future trajectories using a dynamic Bayesian network (DBN) that takes into account all the information relevant to the driver’s decision. Although much progress has been made in applying machine learning to lane change decision problems, lane change decisions under low visibility and mixed traffic conditions have not been considered. In addition, empirical data for learning in such environments are difficult to obtain.
Reinforcement learning not only avoids the use of large amounts of manually marked driving data but also simulates complex lane changes. Therefore, people pay more and more attention to solving lane change decisions by reinforcement learning. Chen et al. [34] designed a hierarchical deep reinforcement learning algorithm to learn lane-changing behavior in dense traffic. Jiang et al. [35] formulated the lane-changing behavior as a POMDP process and proposed a combination of recurrent neural network (RNN) and deep Q network (DQN), which is solved according to the global maximum reward. To overcome the limitations of discrete action space in ordinary reinforcement learning algorithms, Wang et al. [36] established the vehicle lane change behavior under continuous action based on deep deterministic policy gradient (DDPG). Lv et al. [37] developed an improved deep reinforcement learning algorithm based on DDPG with well-defined reward functions for autonomous driving tasks in highway scenarios where an autonomous vehicle merges into two-lane road traffic flow and realizes the lane-changing maneuvers. Kim et al. [38] proposed a new reinforcement learning-based lane change decision-making technique that uses a safety inspection module and augmented data. Wang et al. [39] studied how to learn a harmonious deep reinforcement learning-based lane-changing strategy for autonomous vehicles without Vehicle-to-Everything (V2X) communication support. Ammourah et al. [40] proposed a reinforcement learning-based framework for mandatory lane changing of automated vehicles in a non-cooperative environment that utilized the double deep Q-learning algorithm structure that takes relevant traffic states as input and outputs the optimal actions (policy) for the automated vehicle. He proposed [41] a novel observation adversarial reinforcement learning approach for robust lane change decision-making of autonomous vehicles. Additionally, a constrained observation-robust Markov decision process is presented to model lane change decision-making behaviors of autonomous vehicles under policy constraints and observation uncertainties. In low visibility and mixed traffic, reinforcement learning algorithms can not only adapt to the randomness of drivers but also enable collaboration between CAVs in the absence of communication. Therefore, a reinforcement learning algorithm is applied in this paper to solve the lane change decision problem in this complex environment.
In low-visibility weather, on-board sensors can be disturbed. For example, the functionality of both cameras and LIDAR can be severely degraded in dense fog [42,43]. The sensors can only provide partial information about the observation, which may result in the CAV not obtaining enough information for decision-making. Therefore, it is necessary to consider lane change decisions under partial observation in low-visibility conditions.
In addition, in low visibility, drivers have an increased reaction time and a reduced ability to perceive risks [44,45]. Drivers tend to reduce the headway time distance and shorten the following distance until the view becomes better or they spot the vehicle in front of them [46].
Over the past few years, the rapid arrival of the era of intelligent technology has led to breakthroughs in fuzzy logic technology. Compared to other technologies, fuzzy logic technology has passed through several stages and is more mature, attracting people to research on fuzzy logic technology. Huang et al. [47] distinguished the main differences between interval type 2 (IT2), generalized type 2 (GT2) and interval type 3 (IT3) fuzzy logic systems (FLS) and illustrated the design of IT3-FLS through online identification, offline time series modeling and robot control systems. Jomaa et al. [48] developed a fuzzy logic method to control the actuators that are installed inside the greenhouse for heating, ventilation, humidification and the physical model of the greenhouse used in the Simulink/Matlab environment, which was carefully designed to simulate temperature and indoor humidity. Huang et al. [49] proposed a simple but practical fuzzy logic position tracking controller for a wheeled mobile robot and used particle swarm optimization to turn the parameters of the scaling and membership functions to optimize this controller. Rajagiri et al. [50] designed and developed a fuzzy logic controller in MATLAB Simulink to operate the DC motor. It is experimentally demonstrated that the proposed fuzzy controller results in a better response compared to the normal response of a DC motor. In summary, fuzzy logic technology, as a mature control technique, has broad application prospects and good control performance, especially in the field of intelligent technology where it has been widely applied and researched. With the continuous development and improvement of technology, it is believed that fuzzy logic technology will play its advantages in more fields, such as the research area of mixed traffic flow control of CAV and HDV, and bring more convenience and practical application effects to people.
In summary, the existing research has addressed the lane change decision problem for CAV by various modeling tools and algorithms. However, to our best knowledge, there is no research on the lane change decision for CAV in low-visibility mixed traffic conditions. This paper aims at formulating the lane change decision for CAV as a partially observed problem and solved it using a reinforcement learning algorithm. Numerical experiments are conducted to test the effectiveness and practicality of our algorithm.

3. Problem Formulation

A typical lane change scenario is shown in Figure 1. Lane 1 is the current lane and lane 2 is the target lane. Vehicle A is the CAV preparing to change lanes. Vehicle B and vehicle D are the vehicles ahead of CAV A in the target lane and current lane, respectively, and we do not consider the types of these two vehicles. Vehicle L and vehicle C are the vehicles behind CAV A in the current lane and the target lane, respectively. In fact, the decision timing of CAV A’s lane change is mainly influenced by these two vehicles. This paper focuses on the effect of CAV A lane change when vehicle L and vehicle C are HDVs in low visibility.
In this paper, we assume that the area affected by the lane change is the sensing area of the CAV, i.e., when the CAV changes lanes, the safety judgment will be made for the vehicles in the sensing area. Before describing the problem, the driving behavior of the HDV is modeled: First, we analyzed the influence of low visibility and mixed traffic conditions on drivers. Then, we propose a driver longitudinal control model containing five states under low-visibility mixed traffic conditions. Based on the above, we describe the lane change decision problem to be studied.

3.1. The Longitudinal Control Model for Drivers in Low-Visibility Mixed Traffic Conditions

In the lane change scenario shown in Figure 1, we present the car following states of HDV L and HDV C under low visibility mixing conditions and the reaction states of HDV C when CAV A is changing lanes.

3.1.1. The Influence of Low Visibility and Mixed Traffic Conditions on Drivers

In low visibility, if the driver cannot make a quick and accurate judgment of the surroundings within his or her safe sight distance, the driver will become hesitant about the next driving operation, which will cause a delay in the driver’s reaction. Obviously, the reaction delay of the driver is closely related to the safety sight distance and visibility without considering the driver’s age, personality and gender. In this paper, we use the stopping sight distance d i s 0 to represent the safety sight distance, as shown in Equation (1):
d i s 0 = v e l 0 × t 0 + ( v e l 0 / 3.6 ) 2 2 u m a x
where v e l 0 indicates the current velocity, t 0 indicates the normal reaction time and u m a x indicates the absolute value of the maximum deceleration.
Quantifying the driver’s reaction delay using visibility and safety sight distance in normal weather. To ensure the absolute safety of the vehicle, we express the safety sight distance of the vehicle in low visibility by twice the safety sight distance in normal weather minus the visibility d i s v , as shown in Equation (2):
v e l 0 × t 0 + t d + ( v e l 0 / 3.6 ) 2 2 u m a x 2 d i s 0 d i s v
where t d indicates the driver’s reaction delay due to restricted vision. The left side of the formula indicates the safe driving distance required in the case of a delayed driver reaction. From the above Equation, the smaller the safety sight distance and the greater the visibility, the smaller the driver’s reaction delay.
In mixed traffic conditions, drivers’ trust in CAVs can affect their driving behavior [51]. In low visibility, drivers have fewer targets to refer to, therefore, drivers’ trust in CAVs will be even more important. For example, in low-visibility car-following, drivers tend to follow CAVs at smaller distances when the trust level is high. On the contrary, drivers will treat the behavior of CAVs with caution.
We assume that the trust factor p obeys a standard normal distribution p ~ N   0 , 1 , and β   p denotes the cumulative probability density of p . We let the range of values of p be (−3, 3), which can cover most of the driver aggressiveness. Then we use the distribution function of β   p to represent the degree of trust of drivers in CAV. When the trust factor p is large or small, the probability of the distribution of β   p is small. When p is near 0, the probability of the distribution of β   p is large. Additionally, in this paper, we assume that the CAVs under traffic conditions can be clearly recognized by the drivers.

3.1.2. The Car following States

Considering the reaction delay and the level of confidence, we classify the car following states of HDVs under low-visibility mixed traffic conditions into four types as shown in Figure 2.
State 1: At the current velocity of the vehicle, the safety sight distance is greater than the visibility and there is no other vehicle in front of the HDV within the visibility range. In this state, the driver has a reaction delay. In addition, since the driver cannot see the situation within the safe time distance, the driver tends to look for the vehicle that he can follow ahead in order to get a sense of security [52], therefore, the driver executes a smaller acceleration b 1 , as shown in Equation (3):
a t = b 1 b 1 > 0
where a(t) is the acceleration taken at moment t.
State 2: At the current velocity of the vehicle, the safety sight distance of the vehicle is greater than the visibility and there is a CAV in front of the HDV within the visibility range. The HDV will choose to follow the CAV in this state. But when the state of the CAV changes, the HDV will have a reaction delay. In addition, the driver’s trust in the CAV also affects the following state of the HDV. We use the modified full velocity difference following model (MFVD) to represent the behavior of the HDV in this state, as shown in Equations (4)–(7) [53]:
a i t = K V Δ x i t , v e l i t , p v e l i t + λ Δ v i t
V ( Δ x i t , v e l i t , p ) = v e l l i m 2 [ tanh Δ x i t h e v e l i t ,   p + t a n h h e v e l i t , p ]
h e v e l i t ,   p = 1 β p ( T + τ i ) v e l i t + h c
Δ x i t = x i 1 t x i t
where K ( K > 0 ) is the sensitivity coefficient of headway spacing, V is a reasonable value of speed, a i t is the acceleration taken at moment t of the i-th HDV, λ(λ > 0) is the response coefficient of velocity difference and Δ x i t is the distance between the i-th HDV and the vehicle in front of it. Δ v i t is the velocity difference between the i-th HDV and the vehicle in front of it. h e v e l i t ,   p denotes the following distance expected by the i-th HDV with trust factor p at a velocity of   v e l i t . T is the driver’s normal reaction time, τ i is the driver’s reaction delay and h c is the safe following distance with respect to the vehicle type.
State 3: At the current velocity of the vehicle, the visibility is greater than the safety sight distance, and there is no other vehicle in front of the HDV within the visibility range. In this state, since the driver can see the traffic conditions within the safety sight distance, the driver will execute a smaller acceleration b 2 until the safety sight distance is greater than the visibility or the vehicle ahead is detected within the visibility. We assume that b 2 is greater than b 1 , as shown in Equation (8):
a t = b 2 b 2 > b 1
State 4: At the current velocity of the vehicle, the safety sight distance is less than the visibility, and there is a vehicle in front of the HDV within the visibility range. Similar to state 2, the driver chooses to follow the vehicle ahead, we also use MFVD to represent this; different from state 2 is that the driver has no reaction delay.

3.1.3. The Response Model to Lane Change

In fact, drivers can see the driving conditions in the adjacent lane in the visibility range. This section describes the state of drivers’ reactions to the lane change of vehicles in adjacent lanes. As in Figure 1, HDV C will react to the lane change of CAV A. We refer to Yu et al.’s description of the driver’s lane change response: the more aggressive the driver, the greater the concern for spatial benefits. Conversely, there is a greater concern for their own safety interests, as shown in Equations (9)–(18) [10]:
max f w a , a 0 1 β q × U safety a + β q × U space a + 1 1
a m i n a a m a x
f w a , a 0 = e T l 2 × a a 0 2 w 1 + v + a × T l v d 2 w 2
U safety a = 1 2 S f t = T l S f t = 0
U space a = 1 2 S P t = T l S P t = 0
S f t = 2 T h 1 , t T b 1 ,   T b T h 1 , t T b 1 ,     o t h e r s
S P t = 1 , T h 2 3 2 T h 2 , t 3 + 1 , 3 < T h 2 , t 0 1 , T h 2 > 0  
T h 1 , t = 0 = P A , t = 0 P C , t = 0 v C , t = 0
T h 1 , t = T l = P A , t = T l P C , t = T l v C , t = 0 + a T l ,     P A , t = T l > P C , t = T l P C , t = T l P A , t = T l v A , t = 0 + a A T l ,     o t h e r s
T h 2 , t = P C , t P A , t v A , t ,     P A , t P C , t P C , t P A , t v C , t ,     o t h e r s
The objective function (9) maximizes security benefits and space benefits. Constraint (10) is an acceleration constraint, where a max is the value of the maximum acceleration or minimum deceleration that can be taken, and a min is the value of the minimum acceleration or maximum deceleration that can be taken. f w a , a 0 is the penalty for acceleration and velocity change, as shown in Equation (11). The security benefits U s a f e t y a and space benefits U s p a c e a are calculated as shown in Equations (12) and (13). S f t is the safety factor, as shown in Equation (14). S P t is the space factor, as shown in Equation (15). T l is the time for vehicles in adjacent lanes to change lanes. a and a 0 are the acceleration at the current moment and the next moment, respectively. w 1 and w 2 are the weight values of acceleration change and velocity change, respectively. v d is the desired acceleration. T h 1 is the time headway between vehicle C and vehicle L, including the time headway at the initial moment T h 1 , t = 0 , the time headway at the moment of lane change completion   T h 1 , t = T l and a A is the acceleration of the CAV A, as shown in Equations (16) and (17). T h 2 , t is the evaluation index of the spatial factor, as shown in Equation (18). T b denotes the safe time headway. P A , t and P C , t denote the longitudinal position of vehicle A and vehicle C at moment t, respectively. v A , t and v C , t denote the longitudinal velocity of vehicle A and vehicle C at moment t, respectively. Please refer to the literature [10] for details of the driver response model.

3.1.4. The FSM Model

We compose an FSM model of the driver’s four following states and response states to lane change in low visibility and mixed traffic conditions, as shown in Figure 3. The inputs to the model are whether there is a vehicle ahead in visibility c, whether the safety sight distance is greater than visibility g and whether there is a lane changing vehicle in the adjacent lane in visibility r. When the input is 1, it means certainty, and when the input is 0, it means negation. The output of the model is the five states in the FSM.

3.2. The Lane Change Decision Problem

In this section, we first give the rule-based lane change decision method. In a typical lane change scenario as shown in Figure 1, CAV A usually only needs to consider the safe distance F B A t from vehicle B and vehicle D ahead, as shown in Equation (19):
F B A t = v 2 A t 2 a t v 2 B t 2 a m a x
where v A and v B are the current velocities of vehicle A and vehicle B at moment t, respectively. Considering the driver’s comfort, the deceleration or acceleration value of CAV A should not be too large, as shown in Equation (20):
a t = a m i n + v A t v l i m a m a x a m i n
where v l i m is the limited velocity. The safety distance F D A t between vehicle A and vehicle D is also calculated by Equations (19) and (20).
Within the sensing range of CAV A, if vehicle C is behind the target lane, then CAV A needs to focus on HDV C when changing lanes. There are mainly two cases, one is when the distance between CAV A and HDV C is greater than a specific value   F C A t , as shown in Equation (21). In this case, CAV A can directly change lanes without being affected by HDV C.
F C A t = v C t T + a m a x T 2 2 + v C t + a m a x T 2 2 a m i n v A 2 t 2 a m a x
Another case is the distance between CAV A and HDV C is less than F C A . Under the hard rule, CAV A will drop the lane change. Otherwise, CAV A needs to interact with HDV C and change lanes at the right time according to HDV C’s response. We usually do not need to consider the interference of vehicle L.
In low visibility, although the sensing module will be disturbed, CAV A can still perceive the position of vehicle B, vehicle C and vehicle D most of the time. However, outside of visibility, HDV C cannot find CAV A preparing to change lanes, which will result in no normal interaction between the two vehicles. Therefore, according to Figure 2, we divide the lane change problem of CAV A into four situations based on the relationship between visibility and safety sight distance and the presence or absence of preceding vehicles in visibility, as shown in Figure 4.
Figure 4 shows four lane change scenarios for CAV A. Scenario (a) is a lane change situation where the safety sight distance of HDV C is greater than visibility and vehicle B is not within visibility. Scenario (b) is a lane change situation where the safety sight distance of HDV C is greater than the visibility and the vehicle ahead is visible. Scenario (c) is a lane change situation where the safety sight distance of HDV C is greater than the visibility and vehicle B is not within visibility. Scenario (d) is a lane change situation where the safety sight distance of HDV C is greater than the visibility and vehicle B is not within visibility.
For the lane change in the above special scenarios, the rule-based lane change strategy ensures the absolute safety of the lane change vehicles; therefore, it is still valid. However, it is sometimes too conservative to use a uniform strategy to solve all scenarios; therefore, we try to use reinforcement learning to adopt a corresponding lane change strategy for each scenario.
To show the current problem and the solution method more clearly, we list a schematic diagram, as shown in Figure 5.

4. Solution Methodology

In this section, first, we use POMDP to describe the CAV A lane change problem in Section 3. Then, we develop a modified DDPG algorithm to solve this problem.

4.1. The POMDP Mathematical Model

Subject to extreme environments, CAV perceives an increased probability of missing or incorrect information; therefore, we use POMDP to describe this process. Typically, POMDP consists of a seven-tuple S , A , T , R , Z , O , γ , where X is the set of all environmental states s t at moment t . A is the set of all possible actions a t at moment t .
T is the transfer function; T s t , a t 1 , s t 1 = P ( s t | a t 1 , s t 1 ) denotes the probability of taking action a t 1 in state s t 1 to get the state of moment t as   s t . R is the reward function; r s , a denotes the reward received for taking action a in state s . Z is the set of all possible observation states z t at moment t . O is the observation function; O s t , a t 1 , z t 1 = P ( s t | a t 1 , z t 1 ) denotes the probability of taking action a t 1 to get the state of moment t as s t when the observation value is z t 1 . γ is the discount factor. Based on the POMDP mathematical model, the lane change problem for CAV is described as follows:
  • State S : State of the entire lane change system, including CAV A and surrounding vehicles. The system coordinate is shown in Figure 1, based on which, the state information can be described by the longitudinal position, lateral position and velocity of all vehicles, as shown in Equation (22):
    s t = x A , t , x B , t , x D , t , x C , t , x L , t , y A , t , y B , t , y D , t , y C , t , y L , t , v A , t , v B , t , v D , t , v C , t , v L , t
    where the velocity limit is related to the current visibility of the highway.
  • Action A: The longitudinal acceleration and steering angle of the CAV are used as defined action parameters, as shown in Equation (23):
    a t = u t , δ t
  • Transfer function T : The dynamic model of the lane change system, which is difficult to describe precisely.
  • Observation Z : Due to the influence of the environment or the physical limitations of the sensor itself, CAV A can only obtain the information about other vehicles when sensor noise is present or missing. Z is defined as:
    z t = x A , t , x B , t , x D , t , x C , t , x L , t , y A , t , y B , t , y D , t , y C , t , y L , t , v A , t , v B , t , v D , t , v C , t , v L , t
  • Reward function R : The reward function is the key to achieve the goal of lane change. We design the reward function from the perspective of safety, efficiency and comfort:
    R c o m = α · u x , t ˙ 2 + β · u y , t ˙ 2
    where R c o m denotes the reward function of the comfort level. u x , t ˙ is the longitudinal acceleration rate. u y , t ˙ is the lateral acceleration rate. α and β are the weighting factors of the longitudinal acceleration rate and the lateral acceleration rate, respectively. The comfort reward is introduced mainly to reduce sudden acceleration and deceleration of the CAV A.
In terms of efficiency, the goal is to get the CAV to the centerline of the target lane as quickly as possible without exceeding the velocity limit, as shown in Equations (26)–(29):
R t i m = Δ t
R p o s = y t y *
R s p e = v t v *
R e f f = ω 1 · R t i m + ω 2 · R p o s + ω 3 · R s p e
where R t i m denotes the reward function for the lane change time. Δt denotes the lane change time. R p o s denotes the reward function for position. y t is the current lateral position of the vehicle. y * is the target lane centerline. R s p e denotes the reward function for lane change velocity. v t is the current longitudinal velocity of the vehicle. v * is the limiting velocity at the current visibility. R e f f denotes the reward function for efficiency, and ω 1 , ω 2 , ω 3 are the weights of efficiency, position and velocity, respectively.
Safety is a necessary condition for the vehicle to change lanes. Therefore, when a collision occurs, we give a larger penalty. The safety reward function R s a f is shown in Equation (30):
R s a f = m
where m denotes a larger penalty.
  • Discount factor γ : We use the action value function to evaluate the goodness of the current action, and the value function Q t a r g e t   is updated to consider future rewards with discounts γ , as shown in Equation (31):
    Q t a r g e t = r t + γ s t + 1 S P s t , a t , s t + 1 V s t + 1
    where r t is the reward corresponding to the state s t and action a t at moment t . P s t , a t , s t + 1 is a dynamic model of the system that represents the probability of the system changing from state s t + 1 to state s t after taking action a t .

4.2. The Modified DDPG Algorithm

DDPG is a deep reinforcement learning algorithm that performs continuous control. To obtain better training stability, the algorithm is designed with four networks, including the value network, the strategy network and their respective target networks, as shown in Figure 6.
In Figure 6, θ and θ ¯ are the parameters of the policy network and the target policy network, respectively. W and w ¯ are the parameters of the value network and the target value network, respectively. μ denotes the strategy network.
The decision-making of DDPG is a deterministic behavior. To ensure that the environment is fully explored during training and thus improves training efficiency, we construct random actions by adding an Ornstein–Uhlenbeck (OU) noise with a mean of 0 and a standard deviation of 0.6. The goal of the value network is to continuously update the parameter w , so that its evaluation of taking action a t under state s t is increasingly accurate. The parameter w is updated by timing differences as shown in Equations (32) and (33):
L w = 1 N t = 1 N Y t Q s , a | w | s = s t , a = a t 2
Y t = r t + γ Q s , μ ( s | θ ¯ ) | w ¯ | s = s t + 1 , a = a t + 1
where N is the number of experiences used for the update. Q denotes the value function. Y t is the target value, which is the sum of the reward r t at the current time step and the future reward with discount γ , as shown in Equation (33).
To ensure the stability of Y t during training, we use the target network to calculate the future reward and take a soft update to update the parameters w ¯ and θ ¯ , as shown in Equations (34) and (35):
θ ¯ = τ θ + 1 τ θ ¯
w ¯ = τ w + 1 τ w ¯
where τ is the soft update parameter.
The goal of the strategy network is to make the value network score its strategies as high as possible, and the parameters θ are updated by gradient descent, as shown in Equation (36):
θ J = 1 N i = 1 N a Q s , a | w | s = s t , a = a t θ μ s | θ | s = s t
However, when the randomness of the environment is large, the complexity of reinforcement learning increases, which makes the training process difficult to converge. As shown in Figure 4, the lane change problem is divided into four scenarios based on the relationship between visibility and safe sight distance. In each scenario, the CAV lane-changing strategy needs to learn and complete two tasks, including adapting to different aggressiveness of HDV C and different initial states of CAV A. The combination of these two randomness can complicate the training process.
Therefore, we combine parameter-based transfer learning (TL) with DDPG to split the original complex training task into two subtasks. The first subtask is for CAV A to complete the lane change task for different trust levels of HDV C. Based on the first subtask, the second subtask is for CAV A to learn to complete the lane change at different lane change initial states. In each scenario, the first subtask training converges and then the second subtask learning is performed. Separate training of the tasks will help CAV to complete the lane changing strategy faster. The lane change learning process is shown in Figure 7.

5. Numerical Experiments

5.1. Experiment Design

First, we compared the DDPG and the modified DDPG algorithm so as to verify the effectiveness of the modified algorithm. Second, we tested the modified DDPG algorithm by adjusting the aggressiveness of the HDV driver and the initial lane change state of the CAV A in the four lane change scenarios. Finally, based on some selected metrics, we compared rule-based and DDPG-based lane change algorithms by designing some specific lane change scenarios.
All experiments were performed in MATLAB2021a on a computer equipped with an AMD Ryzen CoreTM16@ 2.90 GHz and 16.0 GB RAM. In addition, the vehicle model used the classical kinematic-based bicycle model. The position of the vehicle is represented by the position of the vehicle’s center of gravity, and the lane width is 3 m. In addition, we assumed that the normal reaction time of the driver is 0.3   s . The goal of the lane change was from the centerline of lane 1 (1.5 m ) to the centerline of lane 2 (4.5 m ).
Based on the typical lane changing scenario in Figure 1, four lane-changing scenarios were designed. The specific parameter settings for scenario 1 are shown in Table 1. When the visibility is 50 m and the initial velocity of HDV C is 18 m / s , the reaction delay is 0.522 s and the safety sight distance is 68.79 m , obtained by Equations (1) and (2). The safety distance between CAV A and HDV B is 10.5 m , obtained by Equations (19) and (20). Scenarios 2, 3 and 4 have the same parameters as Table 1 except for the initial velocity of HDV C and the initial longitudinal position of vehicle B, as shown in Table 2, Table 3 and Table 4. When the initial speed of vehicle C is 14 m/s, its safe sight distance is 37 m, which is less than visibility. Therefore, the driver has no reaction delay.

5.2. An Effectiveness Analysis of the Modified DDPG

As shown in Figure 8, the blue line represents the reward based on the DDPG algorithm, and the red line represents the reward of the modified DDPG algorithm. Obviously, when trained based on the original DDPG algorithm, the CAV learns poorly and only gets stable reward reporting around 450 episodes, while the modified DDPG can converge to a stable reward around 370 episodes.

5.3. Testing of the Modified DDPG Algorithm

In this paper, we mainly focus on whether the modified DDPG can enable CAV A to choose the appropriate lane change opportunity in each specific scenario. Therefore, in this section, we selected several combinations of driver aggression q (trust factor p ) and the initial position x a of CAV, as shown in the Table 5. In addition, after lane change decision, trajectory optimization and tracking are completed based on polynomial and model predictive control, which is not the focus of this paper.
In order to facilitate the discussion during lane change, we subdivide the state of HDV C into the following 13 types.
  • The longitudinal spacing between CAV A and HDV C is greater than the visibility.
    State 1: visibility is less than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 2: visibility is less than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
    State 3: visibility is greater than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 4: visibility is greater than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
  • The longitudinal spacing between CAV A and HDV C is less than the visibility.
    State 5: CAV A is in the lane change state.
    State 6: CAV A has not started lane change, visibility is less than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 7: CAV A has not started lane change, visibility is less than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
    State 8: CAV A has not started lane change, visibility is greater than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 9: CAV A has not started lane change, visibility is greater than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
    State 10: CAV A completes lane change, visibility is less than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 11: CAV A completes lane change, visibility is less than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
    State 12: CAV A completes lane change, visibility is greater than the safe sight distance of HDV C and there is no vehicle ahead in visibility.
    State 13: CAV A completes lane change, visibility is greater than the safe sight distance of HDV C and there is a vehicle ahead in visibility.
Figure 9 shows the results of the lane change test in scenario 1. In (a), (b) and (c), q is 3 and x a is 10 m. In (d), (e) and (f), q is −3, and x a is 10 m. In (g), (h) and (i), q is 3 and x a is 30 m. In (j), (k) and (l), q is 3 and x a is 70 m. As shown in (a), when q is 3, CAV A learns to accelerate during the lane change and also makes itself a sufficiently longitudinal distance from HDV C when the lane change is completed. In contrast, HDV C chooses to yield when q is −3. In this case, considering safety and efficiency, CAV A will change the lane decisively, as shown in (d) and (e). Simultaneously, the longitudinal spacing between CAV A and HDV C is less than the visibility at the end of the lane change, as shown in (c) and (f). When the initial longitudinal position of CAV A is 30 m, CAV A is able to complete the lane change by proper acceleration, as shown in (g) and (h). Due to the acceleration of HDV C during the lane change, the safety sight distance of HDV C is greater than the visibility when CAV A has just completed the lane change, as shown in (i). When the initial longitudinal position of CAV A is 70 m, CAV A has enough space for lane change, even if q of HDV C is 3, and CAV A can complete the lane change quickly, as shown in (j) and (k). At this moment, the longitudinal spacing between HDV C and CAV A has been greater than the visibility during the whole lane change, as shown in (l).
Different from scenario 1, the longitudinal position of HDV B in scenario 2 is 40 m. Figure 10 shows the results of the lane change test in scenario 2. In (a), (b) and (c), q is −3 and x a is 10 m. In (d), (e) and (f), q is 3 and x a is 10 m. In (g), (h) and (i), q is 3 and x a is 30 m. When CAV A has just finished the lane change, HDV C is in state 11 due to the acceleration. Comparing (a) and (b), we can find that CAV A chooses to accelerate and complete the lane change after 50 m when HDV C is more aggressive.
After that, HDV C comes to the car following state shown in Figure 3 and comes to state 13, as shown in (f). When CAV A changes lanes at 30 m, CAV A can complete the lane change by accelerating appropriately. However, different from scenario 1, CAV A needs to maintain the distance from vehicle B. The results of lane change are shown in (g) and (h). In scenarios 3 and 4, the initial velocity of HDV C is smaller. Compared with scenarios 1 and 2, only the initial state of HDV C is different, as shown in Figure 11. While CAV A can easily learn the lane change strategy, the specific lane change test results are not given here.

5.4. Comparison with the Rule-Based Lane Change Decision Algorithm

To further demonstrate the advantages of the modified DDPG algorithm for lane change in this particular environment, this section compares the rule-based lane change decision with the modified DDPG algorithm. In each of the four scenarios, we take any value in the range of (−3, 3) as q of HDV C, and any value in the longitudinal range between HDV C and vehicle B as the initial longitudinal position of CAV A. In each scenario, we conducted 100 test experiments for each decision algorithm and counted the success rate, collision rate and failure rate, as shown in Table 6. In addition, the threshold processing is done for the lane change trajectory learned by the DDPG algorithm so that it can make the wheel steering reduce to 0 after reaching the target lane. The rule-based lane change decision will guarantee the absolute safety of the lane change. When the current conditions cannot achieve the guaranteed safe lane change, CAV A will stay in the current lane.
As shown in Table 6, firstly, the smaller the initial velocity of HDV C, the larger the longitudinal range of the lane change and the higher the success rate. For example, the success rate of scenario 3 can reach 98.5%, while the success rate of Scenario 2 is only 93.1%. Secondly, the success rate of lane change based on the modified DDPG is higher than that of the rule-based in all four scenarios and reaches more than 90%, which indicates that the modified DDPG is well adapted to the randomness of low visibility and mixed traffic conditions. Thirdly, for the collision cases that occur with the modified DDPG, we analyze that they are caused by the incomplete learning of the DDPG in some cases. Additionally, the incomplete cases are due to the fact that CAV A cannot really complete the lane change under the vehicle kinematic constraints in the current stochastic environment. Especially in scenario 2, the incomplete rate reaches 2.3%. The larger initial velocity of HDV C and the smaller initial lane change range increase the learning difficulty of the CAV lane change strategy. Fourth, we can find that the rule-based lane change decision method requires higher lane change conditions. When CAV A cannot achieve the lane change condition described in Section 3.2, CAV A will continue to drive in the current lane. In particular, the lane change failure rate reaches 59.4% for HDV C with an initial velocity of 18 m / s and an initial lane change range of 40   m .
Two special scenarios are selected to illustrate the two decision algorithms and give the historical trajectory, velocity, steering and yaw variation of the vehicle, respectively. In test scenario 1, the q of HDV C is −3, the initial v C is 18 m / s , v A is 14 m / s , the initial longitudinal spacing between HDV C and CAV A is 20 m , the reaction time T is 0.3 s and the absolute values of a m a x and a m i n are 3 m / s 2 .
According to Equation (21), the rule-based lane change decision is executed when the longitudinal spacing between HDV C and CAV A reaches 32.4 m, which is obviously not satisfied. The CAV A stays in the current lane with no change in speed, steering angle and heading angle throughout the process. The trajectory of CAV A is the red line, and the tracks of other vehicles are blue lines, as shown in Figure 12 and Figure 13. However, the modified DDPG finds the lower aggressiveness of HDV C and completes the lane change just at the beginning of the lane change, as shown in Figure 14 and Figure 15. In test scenario 2, the q of HDV C is 3, the initial v C   is 14 m / s and the initial longitudinal spacing between HDV C and CAV A is 10 m . Other conditions are the same as in test scenario 1. Similarly, the rule-based lane change decision will not be executed because the distance condition is not satisfied, as shown in Figure 16 and Figure 17. However, based on the modified DDPG decision algorithm, CAV A accelerates at the beginning. At around 2.7 s, v A approaches 18 m / s , and CAV A begins to change lanes, as shown in Figure 18 and Figure 19.
To better illustrate the effect of the modified DDPG, the emergency lane change scenario and the non-emergency lane change scenario are also selected. The success rate and time for lane change are given, respectively, and compared with both rule-based and DQN-based.
The comparison results are shown in Table 7 and Figure 20. The success rate of the proposed method is 90.36% and the time required is 5.2 s in the emergency lane change scenario, both of which are better than the rule-based 45.12% and 7.6 s and the DQN-based 68.23% and 6.2 s. The comparison results are shown in Table 8 and Figure 21. The success rate of the proposed method is 92.15% and the time required is 5.8 s in the non-emergency lane change scenario. The success rate of the proposed method is better than the rule-based 47.98% and the DQN-based 64.58%, but the time required is only better than the rule-based 7.3 s and slightly weaker than the DQN-based 5.7 s. Overall, our proposed algorithm is clearly superior in both emergency and non-emergency lane change scenarios.

6. Conclusions

In this paper, we present a lane change decision problem for CAVs in low visibility and mixed traffic conditions.
We first consider the effects of low visibility and mixed traffic conditions on the CAV, and then propose an FSM model to describe the driving states of the HDV. Secondly, we use the modified DDPG algorithm to learn the lane change decision of the CAV. Finally, we verify that the modified DDPG converges faster. It is worth noting that the modified DDPG is more adaptive to the environment and can find more lane change opportunities compared to the rule-based lane change decision.
In the future, obtaining the actual driving data of HDV n low visibility and mixed traffic conditions can be further considered. It is also possible to further explore the issues of lane change decisions and overtaking decisions for multiple lanes in low visibility and mixed traffic conditions. On the basis of vehicle micro-control, the traffic operation rule and vehicle scheduling in this environment can continue to be studied.

Author Contributions

Conceptualization, B.G., R.W., T.W. and C.L.; methodology, Z.X. and P.G.; software, R.W.; validation, Z.X. and R.W.; formal analysis, B.G., Z.X. and T.W.; investigation, Z.X.; resources, T.W. and P.G.; data curation, Z.X. and P.G.; writing—original draft preparation, B.G., Z.X. and R.W.; writing—review and editing, Z.X. and R.W.; visualization, B.G., T.W., R.W. and C.L.; supervision, T.W. and C.L.; project administration, T.W., P.G. and C.L. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the Scientific Research Project of the Education Department of Jilin Province (Grant No. JJKH20221020KJ), the Key Laboratory of Transport Industry of Big Data Application Technologies for Comprehensive Transport, Ministry of Transport, China Academy of Transportation Science (Grant No. 2021B1206) and the Qingdao Social Science Planning Research Project (Grant No. 2022-389), who are partly supporting this work.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, G.; Yang, Y.; Zhang, T.; Qu, X.; Cao, D.; Cheng, B.; Li, K. Risk assessment based collision avoidance decision-making for autonomous vehicles in multi-scenarios. Transp. Res. Part C Emerg. Technol. 2021, 122, 102820. [Google Scholar] [CrossRef]
  2. Deng, J.-H.; Feng, H.-H. A multilane cellular automaton multi-attribute lane-changing decision model. Phys. A Stat. Mech. Its Appl. 2019, 529, 121545. [Google Scholar] [CrossRef]
  3. Hruszczak, M.; Lowe, B.T.; Schrodel, F.; Freese, M.; Bajcinca, N. Game Theoretical Decision Making Approach for a Cooperative Lane Change. IFAC-PapersOnLine 2020, 53, 15247–15252. [Google Scholar] [CrossRef]
  4. Das, A.; Khan, M.N.; Ahmed, M.M. Detecting lane change maneuvers using SHRP2 naturalistic driving data: A comparative study machine learning techniques. Accid. Anal. Prev. 2020, 142, 105578. [Google Scholar] [CrossRef]
  5. Xu, T.; Zhang, Z.; Wu, X.; Qi, L.; Han, Y. Recognition of lane-changing behaviour with machine learning methods at freeway off-ramps. Phys. A Stat. Mech. Its Appl. 2021, 567, 125691. [Google Scholar] [CrossRef]
  6. Yoneda, K.; Suganuma, N.; Yanase, R.; Aldibaja, M. Automated driving recognition technologies for adverse weather conditions. IATSS Res. 2019, 43, 253–262. [Google Scholar] [CrossRef]
  7. Furda, A.; Vlacic, L. Multiple Criteria-Based Real-Time Decision Making by Autonomous City Vehicles. IFAC Proc. Vol. 2010, 43, 97–102. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, D.; Srivastava, A.; Ahn, S. Harnessing connected and automated vehicle technologies to control lane changes at freeway merge bottlenecks in mixed traffic. Transp. Res. Part C Emerg. Technol. 2021, 123, 102950. [Google Scholar] [CrossRef]
  9. Karimi, M.; Roncoli, C.; Alecsandru, C.; Papageorgiou, M. Cooperative merging control via trajectory optimization in mixed vehicular traffic. Transp. Res. Part C Emerg. Technol. 2020, 116, 102663. [Google Scholar] [CrossRef]
  10. Yu, H.; Tseng, H.E.; Langari, R. A human-like game theory-based controller for automatic lane changing. Transp. Res. Part C Emerg. Technol. 2018, 88, 140–158. [Google Scholar] [CrossRef]
  11. Chen, K.; Pei, X.; Okuda, H.; Zhu, M.; Guo, X.; Guo, K.; Suzuki, T. A hierarchical hybrid system of integrated longitudinal and lateral control for intelligent vehicles. ISA Trans. 2020, 106, 200–212. [Google Scholar] [CrossRef] [PubMed]
  12. Yi, H.; Edara, P.; Sun, C. Modeling Mandatory Lane Changing Using Bayes Classifier and Decision Trees. IEEE Trans. Intell. Transp. Syst. 2014, 15, 647–655. [Google Scholar] [CrossRef]
  13. Jin, C.J.; Knoop, V.L.; Li, D.W.; Meng, L.Y.; Wang, H. Discretionary lane-changing behavior: Empirical validation for one realistic rule-based model. Transp. A 2019, 15, 244–262. [Google Scholar] [CrossRef] [Green Version]
  14. Xi, C.; Shi, T.; Wu, Y.; Sun, L. Efficient Motion Planning for Automated Lane Change based on Imitation Learning and Mixed-Integer Optimization. In Proceedings of the 23rd IEEE International Conference on Intelligent Transportation Systems, Rhodes, Greece, 20–23 September 2020. [Google Scholar]
  15. Jin, H.; Duan, C.; Liu, Y.; Lu, P. Gauss mixture hidden Markov model to characterise and model discretionary lane-change behaviours for autonomous vehicles. IET Intell. Transp. Syst. 2020, 14, 401–411. [Google Scholar] [CrossRef]
  16. Tang, J.J.; Liu, F.; Zhang, W.H.; Ke, R.M.; Zou, Y.J. Lane-changes prediction based on adaptive fuzzy neural network. Expert Syst. Appl. 2018, 91, 452–463. [Google Scholar] [CrossRef]
  17. Sheikh, M.S.; Wang, J.; Regan, A. A game theory-based controller approach for identifying incidents caused by aberrant lane changing behavior. Phys. A Stat. Mech. Its Appl. 2021, 580, 126162. [Google Scholar] [CrossRef]
  18. Lowe, R.; Wu, Y.; Tamar, A.; Harb, J.; Abbeel, P.; Mordatch, I. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  19. Peng, H.X.; Shen, X.M. Multi-Agent Reinforcement Learning Based Resource Management in MEC- and UAV-Assisted Vehicular Networks. IEEE J. Sel. Areas Commun. 2021, 39, 131–141. [Google Scholar] [CrossRef]
  20. Zheng, Z.D. Recent developments and research needs in modeling lane changing. Transp. Res. Part B—Methodol. 2014, 60, 16–32. [Google Scholar] [CrossRef] [Green Version]
  21. Shalev-Shwartz, S.; Shammah, S.; Shashua, A. On a formal model of safe and scalable self-driving cars. arXiv 2017, arXiv:1708.06374. [Google Scholar]
  22. Gipps, P.G. A model for the structure of lane-changing decisions. Transp. Res. Part B Methodol. 1986, 20, 403–414. [Google Scholar] [CrossRef]
  23. Rickert, M.; Nagel, K.; Schreckenberg, M.; Latour, A. Two lane traffic simulations using cellular automata. Phys. A 1996, 231, 534–550. [Google Scholar] [CrossRef] [Green Version]
  24. Kesting, A.; Treiber, M.; Helbing, D. General lane-changing model MOBIL for car-following models. Transp. Res. Rec. 2007, 1999, 86–94. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, Z.; Zhao, X.M.; Xu, Z.G.; Li, X.P.; Qu, X.B. Modeling and field experiments on autonomous vehicle lane changing with surrounding human-driven vehicles. Comput. Aided Civ. Infrastruct. Eng. 2021, 36, 877–889. [Google Scholar] [CrossRef]
  26. Sun, D.; Kondyli, A. Modeling Vehicle Interactions during Lane-Changing Behavior on Arterial Streets. Comput. Aided Civ. Infrastruct. Eng. 2010, 25, 557–571. [Google Scholar] [CrossRef]
  27. Peng, J.S.; Guo, Y.S.; Fu, R.; Yuan, W.; Wang, C. Multi-parameter prediction of drivers’ lane-changing behaviour with neural network model. Appl. Ergon. 2015, 50, 207–217. [Google Scholar] [CrossRef]
  28. Zhao, C.; Li, Z.H.; Li, L.; Wu, X.B.; Wang, F.Y. A negotiation-based right-of-way assignment strategy to ensure traffic safety and efficiency in lane changes. IET Intell. Transp. Syst. 2021, 15, 1345–1358. [Google Scholar] [CrossRef]
  29. Hidas, P. Modelling lane changing and merging in microscopic traffic simulation. Transp. Res. Part C Emerg. Technol. 2002, 10, 351–371. [Google Scholar] [CrossRef]
  30. Hidas, P. Modelling vehicle interactions in microscopic simulation of merging and weaving. Transp. Res. Part C Emerg. Technol. 2005, 13, 37–62. [Google Scholar] [CrossRef]
  31. Yoo, J.H.; Langari, R. Stackelberg Game Based Model of Highway Driving. In Proceedings of the 5th Annual Dynamic Systems and Control Division Conference/11th JSME Motion and Vibration Conference, Fort Lauderdale, FL, USA, 17–19 October 2012; pp. 499–508. [Google Scholar]
  32. Chen, T.Y.; Shi, X.P.; Wong, Y.D. Key feature selection and risk prediction for lane-changing behaviors based on vehicles’ trajectory data. Accid. Anal. Prev. 2019, 129, 156–169. [Google Scholar] [CrossRef]
  33. Gindele, T.; Brechtel, S.; Dillmann, R. A probabilistic model for estimating driver behaviors and vehicle trajectories in traffic environments. In Proceedings of the 13th International IEEE Conference on Intelligent Transportation Systems, Funchal, Portugal, 19–22 September 2010; pp. 1625–1631. [Google Scholar]
  34. Chen, Y.L.; Dong, C.Y.; Palanisamy, P.; Mudalige, P.; Muelling, K.; Dolan, J.M. Attention-based Hierarchical Deep Reinforcement Learning for Lane Change Behaviors in Autonomous Driving. In Proceedings of the 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1326–1334. [Google Scholar]
  35. Jiang, S.H.; Chen, J.Y.; Shen, M.C. An Interactive Lane Change Decision Making Model With Deep Reinforcement Learning. In Proceedings of the 7th IEEE International Conference on Control, Mechatronics and Automation (ICCMA), Delft, The Netherlands, 6–8 November 2019; pp. 370–376. [Google Scholar]
  36. Wang, P.; Li, H.H.; Chan, C.Y. Continuous Control for Automated Lane Change Behavior Based on Deep Deterministic Policy Gradient Algorithm. In Proceedings of the 30th IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 1454–1460. [Google Scholar]
  37. Lv, K.; Pei, X.; Chen, C.; Xu, J. A Safe and Efficient Lane Change Decision-Making Strategy of Autonomous Driving Based on Deep Reinforcement Learning. Mathematics 2022, 10, 1551. [Google Scholar] [CrossRef]
  38. Kim, M.-S.; Eoh, G.; Park, T. Reinforcement learning with data augmentation for lane change decision-making. J. Inst. Control Robot. Syst. 2021, 27, 572–577. [Google Scholar] [CrossRef]
  39. Wang, G.; Hu, J.; Li, Z.; Li, L. Harmonious Lane Changing via Deep Reinforcement Learning. IEEE Trans. Intell. Transp. Syst. 2022, 23, 4642–4650. [Google Scholar] [CrossRef]
  40. Ammourah, R.; Talebpour, A. Deep Reinforcement Learning Approach for Automated Vehicle Mandatory Lane Changing. Transp. Res. Rec. 2022, 2677, 712–724. [Google Scholar] [CrossRef]
  41. He, X.; Yang, H.; Hu, Z.; Lv, C. Robust Lane Change Decision Making for Autonomous Vehicles: An Observation Adversarial Reinforcement Learning Approach. IEEE Trans. Intell. Veh. 2022, 8, 184–193. [Google Scholar] [CrossRef]
  42. Wallace, A.M.; Halimi, A.; Buller, G.S. Full Waveform LiDAR for Adverse Weather Conditions. IEEE Trans. Veh. Technol. 2020, 69, 7064–7077. [Google Scholar] [CrossRef]
  43. Heinzler, R.; Schindler, P.; Seekircher, J.; Ritter, W.; Stork, W.; IEEE. Weather Influence and Classification with Automotive Lidar Sensors. In Proceedings of the 30th IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 1527–1534. [Google Scholar]
  44. Chen, C.; Zhao, X.; Liu, H.; Ren, G.; Liu, X. Influence of adverse weather on drivers’ perceived risk during car following based on driving simulations. J. Mod. Transp. 2019, 27, 282–292. [Google Scholar] [CrossRef] [Green Version]
  45. Caro, S.; Cavallo, V.; Marendaz, C.; Boer, E.R.; Vienne, F. Can headway reduction in fog be explained by impaired perception of relative motion? Hum. Factors 2009, 51, 378–392. [Google Scholar] [CrossRef]
  46. Saffarian, M.; Happee, R.; de Winter, J.C.F. Why do drivers maintain short headways in fog? A driving-simulator study evaluating feeling of risk and lateral control during automated and manual car following. Ergonomics 2012, 55, 971–985. [Google Scholar] [CrossRef]
  47. Huang, H.; Xu, H.; Chen, F.; Zhang, C.; Mohammadzadeh, A. An Applied Type-3 Fuzzy Logic System: Practical Matlab Simulink and M-Files for Robotic, Control, and Modeling Applications. Symmetry 2023, 15, 475. [Google Scholar] [CrossRef]
  48. Jomaa, M.; Abbes, M.; Tadeo, F.; Mami, A. Greenhouse Modeling, Validation and Climate Control based on Fuzzy Logic. Eng. Technol. Appl. Sci. Res. 2019, 9, 4405–4410. [Google Scholar] [CrossRef]
  49. Huang, C.; Farooq, U.; Liu, H.; Gu, J.; Luo, J. A Pso-Tuned Fuzzy Logic System for Position Tracking of Mobile Robot. Int. J. Robot. Autom. 2019, 34, 84–94. [Google Scholar] [CrossRef]
  50. Rajagiri, A.K.; Mn, S.R.; Nawaz, S.S.; Suresh Kumar, T. Speed control of DC motor using fuzzy logic controller by PCI 6221 with MATLAB (Conference Paper). E3S Web Conf. 2019, 87, 01004. [Google Scholar] [CrossRef] [Green Version]
  51. Zhao, X.; Wang, Z.; Xu, Z.; Wang, Y.; Li, X.; Qu, X. Field experiments on longitudinal characteristics of human driver behavior following an autonomous vehicle. Transp. Res. Part C Emerg. Technol. 2020, 114, 205–224. [Google Scholar] [CrossRef]
  52. Gao, K.; Tu, H.; Shi, H.; Li, Z. Effect of low-visibility in haze weather condition on longitudinal driving behavior in different car following stages. J. Jilin Univ. 2017, 47, 1716–1727. [Google Scholar]
  53. Jiang, R.; Wu, Q.; Zhu, Z. Full velocity difference model for a car-following theory. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 2001, 64, 017101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. The typical lane change scenario.
Figure 1. The typical lane change scenario.
Mathematics 11 01556 g001
Figure 2. The four car-following states of HDVs under mixed driving conditions in low visibility.
Figure 2. The four car-following states of HDVs under mixed driving conditions in low visibility.
Mathematics 11 01556 g002
Figure 3. The FSM model of HDV in low visibility and mixed traffic conditions.
Figure 3. The FSM model of HDV in low visibility and mixed traffic conditions.
Mathematics 11 01556 g003
Figure 4. The four lane change scenarios for CAV A in low visibility and mixed traffic conditions.
Figure 4. The four lane change scenarios for CAV A in low visibility and mixed traffic conditions.
Mathematics 11 01556 g004
Figure 5. The flow diagram of the problem formulation and methodological logic framework of this paper.
Figure 5. The flow diagram of the problem formulation and methodological logic framework of this paper.
Mathematics 11 01556 g005
Figure 6. The framework of DDPG algorithm.
Figure 6. The framework of DDPG algorithm.
Mathematics 11 01556 g006
Figure 7. The learning process of lane changing strategy based on DDPG and TL.
Figure 7. The learning process of lane changing strategy based on DDPG and TL.
Mathematics 11 01556 g007
Figure 8. Reward changes based on DDPG algorithm and modified DDPG algorithm in scenario 1.
Figure 8. Reward changes based on DDPG algorithm and modified DDPG algorithm in scenario 1.
Mathematics 11 01556 g008
Figure 9. Test results in scenario 1.
Figure 9. Test results in scenario 1.
Mathematics 11 01556 g009
Figure 10. Test results in scenario 2.
Figure 10. Test results in scenario 2.
Mathematics 11 01556 g010
Figure 11. Test results of the state of HDV C in scenario 3 and scenario 4.
Figure 11. Test results of the state of HDV C in scenario 3 and scenario 4.
Mathematics 11 01556 g011
Figure 12. The result of the trajectories of test scenario 1 based on the rule-based lane change decision.
Figure 12. The result of the trajectories of test scenario 1 based on the rule-based lane change decision.
Mathematics 11 01556 g012
Figure 13. The velocity, steering and yaw of test scenario 1 based on the rule-based lane change decision.
Figure 13. The velocity, steering and yaw of test scenario 1 based on the rule-based lane change decision.
Mathematics 11 01556 g013
Figure 14. The result of the trajectory of test scenario 1 based on the modified DDPG.
Figure 14. The result of the trajectory of test scenario 1 based on the modified DDPG.
Mathematics 11 01556 g014
Figure 15. The velocity, steering and yaw of test scenario 1 based on the modified DDPG.
Figure 15. The velocity, steering and yaw of test scenario 1 based on the modified DDPG.
Mathematics 11 01556 g015
Figure 16. The result of the trajectory of test scenario 2 based on the rule-based lane change decision.
Figure 16. The result of the trajectory of test scenario 2 based on the rule-based lane change decision.
Mathematics 11 01556 g016
Figure 17. The velocity, steering and yaw of test scenario 2 based on the rule-based lane change decision.
Figure 17. The velocity, steering and yaw of test scenario 2 based on the rule-based lane change decision.
Mathematics 11 01556 g017
Figure 18. The result of the trajectory of test scenario 2 based on the modified DDPG.
Figure 18. The result of the trajectory of test scenario 2 based on the modified DDPG.
Mathematics 11 01556 g018
Figure 19. The velocity, steering and yaw of test scenario 2 based on the modified DDPG.
Figure 19. The velocity, steering and yaw of test scenario 2 based on the modified DDPG.
Mathematics 11 01556 g019
Figure 20. The success rate and time for lane change of the emergency lane change scenario based on three methods.
Figure 20. The success rate and time for lane change of the emergency lane change scenario based on three methods.
Mathematics 11 01556 g020
Figure 21. The success rate and time for lane change of the non-emergency lane change scenario based on three methods.
Figure 21. The success rate and time for lane change of the non-emergency lane change scenario based on three methods.
Mathematics 11 01556 g021
Table 1. Parameter setting for scenario 1.
Table 1. Parameter setting for scenario 1.
ParameterValueParameterValue
Visibility V
Initial velocity of CAV A v a
50   m
14 m / s
Limit velocity v l i m
Initial velocity of HDV C v c
16 m / s
18 m / s
Initial velocity of HDV B v b
Initial longitudinal position of HDV B x b
Initial longitudinal position of HDV C x c
Initial lateral position of HDV B y b
Initial lateral position of HDV D y d
* Aggression factor of HDV C q
14 m / s
100 m
0 m
4.5 m
1.5 m
−3~3
Initial velocity of HDV D v d
Initial longitudinal position of HDV D x d
Initial lateral position of CAV A y a
Initial lateral position of HDV C y c
Initial longitudinal position of CAV A x a
14 m / s
100 m
1.5 m
4.5 m
15~85 m
*   We assume that the driver’s trust factor p and aggression factor q are equal.
Table 2. Parameter setting for Scenario 2.
Table 2. Parameter setting for Scenario 2.
ParameterValueParameterValue
Initial   velocity   of   HDV   C   v c 18   m / s Initial   longitudinal   position   of   HDV   B   x b 40   m
Table 3. Parameter setting for Scenario 3.
Table 3. Parameter setting for Scenario 3.
ParameterValueParameterValue
Initial   velocity   of   HDV   C   v c 14   m / s Initial   longitudinal   position   of   HDV   B   x b 100   m
Table 4. Parameter setting for Scenario 4.
Table 4. Parameter setting for Scenario 4.
ParameterValueParameterValue
Initial   velocity   of   HDV   C   v c 14   m / s Initial   longitudinal   position   of   HDV   B   x b 40   m
Table 5. Test combination of q and xa.
Table 5. Test combination of q and xa.
x a ( m )
q 103070
−3(−3, 10)\\
3(3, 10)(3, 30)(3, 70)
Table 6. Performances of the modified DDPG and the rule-based lane change decision algorithms.
Table 6. Performances of the modified DDPG and the rule-based lane change decision algorithms.
ScenarioThe Modified DDPGThe Rule-Based
SuccessCollisionUnfinishedSuccessUnfinished
Scenario 197.2%1.8%1%85.4%14.6%
Scenario 293.1%4.6%2.3%40.6%59.4%
Scenario 398.5%0.4%1.1%91.1%8.9%
Scenario 495.8%2.6%1.6%68.9%31.1%
Table 7. Comparison of modified DDPG and rule-based and DQN-based in the emergency lane change scenario.
Table 7. Comparison of modified DDPG and rule-based and DQN-based in the emergency lane change scenario.
Success Rate for Lane ChangeTime for Lane Change
Our Method90.36%5.2 s
Rule-based45.12%7.6 s
DQN-based68.23%6.2 s
Table 8. Comparison of modified DDPG and rule-based and DQN-based in the non-emergency lane change scenario.
Table 8. Comparison of modified DDPG and rule-based and DQN-based in the non-emergency lane change scenario.
Success Rate for Lane ChangeTime for Lane Change
Our Method92.15%5.8 s
Rule-based47.98%7.3 s
DQN-based64.58%5.7 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, B.; Xu, Z.; Wei, R.; Wang, T.; Lin, C.; Gao, P. Reinforcement Learning-Based Lane Change Decision for CAVs in Mixed Traffic Flow under Low Visibility Conditions. Mathematics 2023, 11, 1556. https://doi.org/10.3390/math11061556

AMA Style

Gong B, Xu Z, Wei R, Wang T, Lin C, Gao P. Reinforcement Learning-Based Lane Change Decision for CAVs in Mixed Traffic Flow under Low Visibility Conditions. Mathematics. 2023; 11(6):1556. https://doi.org/10.3390/math11061556

Chicago/Turabian Style

Gong, Bowen, Zhipeng Xu, Ruixin Wei, Tao Wang, Ciyun Lin, and Peng Gao. 2023. "Reinforcement Learning-Based Lane Change Decision for CAVs in Mixed Traffic Flow under Low Visibility Conditions" Mathematics 11, no. 6: 1556. https://doi.org/10.3390/math11061556

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop