Next Article in Journal
MOFs-Derived Nano-CuO Modified Electrode as a Sensor for Determination of Hydrazine Hydrate in Aqueous Medium
Next Article in Special Issue
Accurate and Practical Energy Detection over α-μ Fading Channels
Previous Article in Journal
Fast Method of Registration for 3D RGB Point Cloud with Improved Four Initial Point Pairs Algorithm
Previous Article in Special Issue
Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters in Infrared and Visible Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Traffic Signal Control with Traffic Flow Prediction in Multi-Intersection

Department of Software, Gachon University, Gyeonggi 13120, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(1), 137; https://doi.org/10.3390/s20010137
Submission received: 5 November 2019 / Revised: 17 December 2019 / Accepted: 17 December 2019 / Published: 24 December 2019
(This article belongs to the Special Issue Intelligent Systems in Sensor Networks and Internet of Things)

Abstract

:
As traffic congestion in cities becomes serious, intelligent traffic signal control has been actively studied. Deep Q-Network (DQN), a representative deep reinforcement learning algorithm, is applied to various domains from fully-observable game environment to traffic signal control. Due to the effective performance of DQN, deep reinforcement learning has improved speeds and various DQN extensions have been introduced. However, most traffic signal control researches were performed at a single intersection, and because of the use of virtual simulators, there are limitations that do not take into account variables that affect actual traffic conditions. In this paper, we propose a cooperative traffic signal control with traffic flow prediction (TFP-CTSC) for a multi-intersection. A traffic flow prediction model predicts future traffic state and considers the variables that affect actual traffic conditions. In addition, for cooperative traffic signal control in multi-intersection, each intersection is modeled as an agent, and each agent is trained to take best action by receiving traffic states from the road environment. To deal with multi-intersection efficiently, agents share their traffic information with other adjacent intersections. In the experiment, TFP-CTSC is compared with existing traffic signal control algorithms in a 4 × 4 intersection environment. We verify our traffic flow prediction and cooperative method.

1. Introduction

Recently, traffic congestion has become a serious problem in most cities. Since the capacity of the road is limited, it is difficult to handle the increasing traffic flow. Traffic signal control [1,2,3] and traffic flow prediction [4,5,6] methods are the most effective solution to mitigate traffic congestion. In general, traffic signal control has the advantage of being able to cope immediately with the amount of traffic flow occurring at the intersection level [7]. On the other hand, traffic flow prediction can cover larger regions based on historical traffic data [8]. The adaptive traffic signal control [9,10] is one of the most effective ways to improve the road environment. Adaptive traffic signal control takes the action of switching the traffic phase or adjusting the traffic signal length.
Reinforcement learning is widely used in traffic signal control research to find the optimal solution [11,12]. Reinforcement learning agent learns optimal policy by exploring uncertain environment. In traffic signal control, the agent gets a state from the road such as traffic flow and takes action based on policy. The agent then receives a reward and learns to maximize the discounted cumulative reward. However, such an algorithm has a limitation that state and action space will grow exponentially as the environment size increases [13]. Due to this limitation, reinforcement learning is difficult to apply in real-world situations.
In recent years, deep reinforcement learning, which combines traditional reinforcement learning and deep neural network, was introduced. This algorithm has attracted a lot of attention because of deep Q-network (DQN) [13] applied to AlphaGo [14]. Deep reinforcement learning was first applied to game environments such as Go and Tetris, and recently to adaptive traffic signal control research. Due to high performance of DQN, many researches of traffic signal control have been implemented by DQN [15,16]. However, there is less research on deep reinforcement learning for cooperative traffic signal control at multi-intersections and a lack of possibility of applying in the real world.
In this paper, we propose the cooperative traffic signal control with traffic flow prediction for a multi-intersection environment. The main contributions of our research are as follows:
  • Cooperative traffic signal control method for harmonious multi-intersection environment.
  • Traffic flow prediction for the real world.
For cooperative traffic signal control at a multi-intersection [17,18], we consider all adjacent intersections and derive a global optimal value. This method can help to derive the global optimal Q-value by estimating the probability of incoming flow from other intersections. Traffic flow prediction may predict the future traffic flow in a different way. Most existing traffic signal control studies are trained in virtual simulation [19]. In the real world, traffic flow is especially heavily influenced by variables such as weather, time, and day [20,21]. However, it is difficult to consider these variables in the virtual simulator. Thus, we propose the traffic signal control with traffic flow prediction which approaches a real environment.
The remainder of this paper is organized as follows: The related work is presented in Section 2. The background of deep reinforcement learning algorithm and DQN extensions are presented in Section 3. The proposed cooperative methods and traffic flow prediction at multi-intersection are introduced in Section 4. Our algorithm is evaluated in various ways in Section 5. Our conclusion and future work are presented in Section 6.

2. Related Work

Currently, most traffic signal systems have a static traffic signal length, and static traffic signal systems do not flexibly cope with traffic changes in real-time [22]. In order to deal with these limitations, the study of adaptive traffic signal control [9,10] has been actively studied. Adaptive traffic control uses a method to switch the traffic signal phase in real-time based on the traffic volume observed at the intersection. In other words, traffic congestion can be alleviated by giving a green signal length to a lane with more traffic flow.
Various machine learning algorithms are used for adaptive traffic control [3,7,10]. Recently, many fuzzy logic-based algorithms [23,24] have been studied as machine learning algorithms. Fuzzy logic-based adaptive traffic control studies [25,26,27] generally improve the performance of existing studies. Arif A. [28] introduced an adaptive traffic signal control method using image processing techniques. This employ a method of extracting a region of interest (ROI) by recognizing cars on a road area in the image using a four-way camera installed at an intersection and performing work according to a predefined flowchart. However, the adaptive traffic signal control method also has various problems.
Traditional reinforcement learning methods are also widely used for traffic signal control. Reinforcement learning aims to find the optimal policy that maximizes the reward. In reinforcement learning, an agent explores by taking various actions based on the state and learns to maximize the reward by receiving the result of the action as a reward. YK Chin’s study [29] introduces a method for optimizing pre-defined traffic signal phases in a multi-intersection environment using the Q-learning algorithm. It describes how to optimize traffic signal control based on a total of 256 possible traffic conditions, which are categorized into four signal phases and four levels of vehicle queues, respectively, in a traffic network with two intersections. In addition, Hua W. and certain researches [16,30,31] point out the limitations of the hand-crafted rules of the current traffic signal system. They also introduced reinforcement learning, value-based deep reinforcement learning, policy gradient techniques to dynamically adjust traffic signal in response to real-time traffic congestion.
Reinforcement learning basically selects the optimal action based on Q learning, which is not free from the “curse of dimensionality”. Therefore, deep reinforcement learning that combines reinforcement learning and an artificial neural network is proposed [13].
Deep reinforcement learning has received a lot of attention in solving problems such as the “curse of dimensionality”. DQN showed effective performance by approximating high action space through the neural network. Due to the high performance of DQN, deep reinforcement learning has been applied to adaptive traffic signal control [32,33]. It is especially possible to utilize all the state information of intersection through CNN [32]. Since DQN, various extensions such as Double DQN [34] and Dueling DQN [35] have been introduced quickly, and the research [36] combining these extensions improved the learning stability much more than the existing DQN.
Recently, multiple intersection traffic signal control has been studied [18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37], but there is less research of traffic signal control at multiple intersections. Ref [16] proposes an A3C technique with a decentralized coordinated algorithm for multi-intersection traffic signal control. CNN is used to extract vehicle position and velocity information from a snapshot of the simulator. However, most snapshots are inefficient because they do not include the vehicle, resulting in sparse matrix. Ref [37] applied DQN as a traffic signal control algorithm and proposes Q-value transfer for harmonious multiple intersections control. However, several DQN extensions have already appeared that improve on the shortcomings of DQN, and in many areas have been proven to outperform existing DQN. In addition, because the experimental environment is less than 10 intersections, it is difficult to apply in real-world situations. In addition, the existing traffic signal control studies were conducted using a virtual simulator, which does not deal with variables such as temperature and weather that actually affect traffic conditions.
In this paper, we propose an algorithm that combines cooperative traffic signal control at multiple intersections and further predicts traffic volume taking into account real-world variables. For cooperative traffic control at multiple intersections, we consider traffic information at adjacent intersections. We estimate the global optimal q-value at multiple intersections by sharing the local optimal q-values of each adjacent intersection. We also consider the various variables that affect actual traffic, thus enabling more advanced training. In an experiment, we demonstrate high performance by applying algorithms combined with various DQN extensions to traffic signal control.

3. Deep Reinforcement Learning and Extensions to DQN

In reinforcement learning, the agent learns an optimal policy by interacting with the environment. The learning process of reinforcement learning is illustrated in Figure 1. At each step, the agent receives the state from an uncertain environment. Then the agent selects the action based on policy. The agent then receives a reward from the environment. The goal is to maximize the cumulative reward. The Q-function predicts the expected future reward given state and action. The Q-function is formulated as follows:
Q π ( s , a ) = E   [ k = 0 γ k r t + k   |   s t = s ,   a t = a ,   π ]
where γ is a discount factor and is to multiply the expected reward, it means that set the lower reward value received at later.

3.1. Deep Reinforcement Learning

In reinforcement learning, it is important that how to define state, action, and reward function, respectively. The state, action, and reward are defined as follows.

3.1.1. State

The state represents the observation information from environment. Since the DQN uses the deep neural network, which is a non-linear approximator, many researchers define snapshot of simulator as state [32,33]. The snapshot of simulator is useful to get the vehicle status such as speed and position of vehicles. However, it needs additional state encoding process to get the traffic state space. In this paper, we simply set the number of vehicles of each lane as state space. Figure 2 shows the example of an intersection. Each intersection has four-arms (edges) and each edge has two lanes. To deal with multi-intersection scenario, we must get the number of vehicles in each lane at each time step.

3.1.2. Action

At each time step, the agent selects the action from given state. In this paper, we define a set of actions A, A = {NSG, EWG, NSLG, EWLG}. Each action of set represents the traffic signal phase and N, S, E, W, L, G represents North, South, East, West, Left, Green and where NSG indicates North-South Green, EWG indicates East-West Green, NSLG indicates North-South Left Green and EWLG indicates East-West Left Green. The example of a traffic signal system is illustrated in Table 1 and the traffic signal system of all intersections are the same. Table 1 shows traffic signal phases for each action {NSG, EWG, NSLG, EWLG}. There are 8 phases and each phase consists of traffic signal such as G(Green), y(yellow) and r(red) in the order R(Right turn lane), S(Straight lane) and L(Left turn lane). For example, phase 0 means that turn on the green signal to north and south, straight and right lanes. Also, each traffic signal phase has a duration. In this paper, we select the traffic signal phase based on observed state. For example, in Figure 2, the number of vehicles in the north and south is greater than in east and west. Thus, if the NSG (= phase 0) phase is already set, keep that phase and increase the duration, otherwise switch to NSG phase and increase the duration. In addition, in order to consider inherent influence between intersections, we propose a cooperative method that shares traffic conditions between intersections. Figure 5 shows an example of our cooperative method. If the traffic signal at intersection 3 is EWG (East-West Green) and the traffic signal at intersection 8 is NSG (North-South Green), intersection 4 is naturally affected by the traffic coming from intersections 3 and 8. To take this effect into account, we propose a cooperative method of transferring the state and action of adjacent intersections to other intersections.

3.1.3. Reward

The reward function is the most important element that differentiates other machine learning algorithms. The reward function is used to evaluate the model performance. Since the goal of reinforcement learning is to maximize the reward, defining reward function is important. In this paper, we define reward function as the changes of average waiting time between time step t and time step t+1. The lower the waiting time, the higher reward. The reward function r is formulated as follows:
R w a i t i n g _ t i m e = W t   W t + 1

3.2. Extensions to DQN

Deep reinforcement learning has achieved remarkable success in many domains from game environment to traffic signal control. DQN [13] is representative deep reinforcement learning algorithm and shows the high performance. Due to the encouraging performance of DQN, various extensions of DQN [34,35] have been introduced. Recently, [36] proposed the combination of extensions and demonstrate high performance in the Atari game environment. They combine 6 extensions with DQN which are Double Q-learning, Dueling network, Prioritized replay [38], Multi-step learning [39], Distributional RL [40] and Noisy Nets [41]. We applied that those algorithms to cooperative traffic signal control and we explain how they are applied to our algorithm.
Double Q-learning improved the DQN by decoupling the selection of the action from its evaluation. This algorithm improved an overestimation problem of conventional Q-learning. In our model, we construct the separate target network from main network. Dueling network is a neural network architecture designed for value-based reinforcement learning. This algorithm divides neural network into value and advantage stream. That is, the Q function of DQN is divided into state value functions and advantage function. This algorithm can get faster learning speed because the advantage function only focuses on the action value. We applied dueling network when constructing neural network in our model. Prioritized replay improves data efficiency, by replaying more often transitions from which there is more to learn. The priority of transition is calculated based on TD (Temporal Difference) error. At each training step, we calculate the priorities of transitions and sampled the high priority transitions. Multi-step learning truncates the n-step rewards, not the single reward. Instead of computing the reward function at every training step, we compute the reward n-steps after. This algorithm is able to minimize the loss and lead to faster learning. Distributional RL is also related to reward. DQN and other algorithms commonly define average waiting time or delay as reward. Thus, reward value is scalar. Instead of a scalar reward, we learn to approximate the distribution of the reward. This algorithm is useful when the future reward is complex and has multimodal features by representing the reward in distribution. NoisyNet proposes a noisy linear later that combines a deterministic and noisy stream. Over time, the network can learn to ignore the noisy stream, but will do so at different rates in different parts of the state space, allowing state conditional exploration with a form of self-annealing. NoisyNet is useful when action space is too large and replaces the existing exploring policy, є-greedy. Algorithm 1 shows a description of the whole algorithm and how extensions of DQN such as Double DQN, Dueling DQN, Prioritized experience replay, Distributional RL, Multi-step learning, and NoisyNet were harmoniously applied to traffic signal control.
Algorithm 1. Cooperative traffic signal control.
Initialize network parameter θ
Initialize target network parameter θ
Initialize replay memory m with size M, batch size B, target network replacement frequency T
Initialize initial state s
for epochs e {1, … N} do
 Observe current state s t
 Select an action a from noisy network
 Get reward r and next state s t + 1
 If |m| > M then
  Delete oldest transition t from m
 end
 Add transition t = ( s t , a, r, s t + 1 ) to replay memory m
s t s t + 1
 If |m| > B then
  Sample B transitions from m based on prioritized experience replay
  Compute multi-step loss based on distributional loss
  Update network θ using loss
  If e == T then
   Update target network θ θ
  end
 end
end

4. TFP-CTSC (Cooperative Traffic Signal Control with Traffic Flow Prediction)

In this section, we propose the cooperative traffic signal control with traffic flow prediction (TFP-CTSC) in a multi-intersection. There are two reasons why we propose our paper: (1) There is less research on traffic signal control in multi-intersections. (2) Existing studies of traffic signal control are dependent on virtual simulation environment. To efficiently improve the performance in multi-intersection environment, we consider the adjacent intersections at the last time step. Also, we complement the limitations of virtual simulation by considering realistic variables that affect traffic flow. Overall architecture of our paper is illustrated in Figure 3. At each time stept, the agent receives the state from environment. At the same time, agent send state to traffic flow prediction model. In traffic flow prediction model, they predict the future traffic flow (statet+1) using realistic variable. And then, return the statet+1 to agent. Agent compute the optimal Q-value from both observed state and predicted future traffic flow to choose the action. For a state and action given, the reward is received. The main goal of our model is also to find the optimal policy which maximize the future reward.

4.1. Traffic Flow Prediction for Real-World

In the real world, traffic flow can be easily affected by those variables such as weather, day, and time. For example, generally, traffic congestion usually occurs more often at rush hour and in weather such as snow or rain [21]. In a simulation environment, it is difficult to deal with these factors. That’s why we construct separate traffic flow prediction model to reflect real-world variables. We select the day, time, and temperature that most influence the traffic flow, and the correlation matrix between each variable is shown in Figure 4a. We can see that highly correlated variable with t+1-flow are t-flow, temperature, and time.
Traffic flow prediction research has same objective as traffic signal control to mitigate the traffic congestion by predicting the future traffic. Our model is trained by using LSTM (Long-Short Term memory) [42] to predict. LSTM is useful to training the time-series data because of a memory layer.
When the state t is received from environment, traffic flow prediction model predicts the state t+1 at time step t+1. The predicted state t+1 is used to optimize the Q-value. We use RMSE (Root Mean Squared Error) as loss function. The loss is 12.9 and RMSE is 0.023. The Q-values are updated as follows:
s t + 1 = f t ( t i m e ,   d a y ,   t e m p e r a t u r e ,   S t )  
a = a r g m a x Q t i ( s t + 1 , a )
Q t + 1 i ( s t i , a t i ) =   Q t i ( s t i , a t i ; θ i ) + η ( R t + 1 +   γ m a x Q t i ( s t + 1 , a ; θ i ) Q t i ( s t i , a t i ; θ i ) )
here s t + 1 is predicted traffic flow of traffic flow prediction model at time step t, a is indices of maximum value in Q t i ( s t + 1 , a ) , θ i and θ i are the parameters of evaluation network and target network, respectively. Our updated Q-value replaces s t + 1 with predicted traffic value of traffic flow prediction model. This newly updated Q-value is getting closer to real-world by traffic flow prediction. The Figure 4b shows the visualized results of traffic flow prediction. The predicted traffic flow is represented as a circle. The closer the circle is to red, the higher traffic flow and the closer to green, the lower traffic flow.

4.2. TFP-CTSC in Multi-Intersection

For cooperative traffic signal control in multi-intersection, we propose transfer planning [43] based cooperative traffic signal control with traffic flow prediction. We model a multi-intersection environment as a multi-agent. Each agent is trained to find local optimal Q-value, and then we try to find global optimal Q-value by transferring the messages which contain local optimal value of each agent. Cooperative traffic signal control is possible by considering not only its own Q-value, but also the Q-value of adjacent intersections. Finally, we can draw the global optimal Q-value of each agent and this Q-value is updated as follows:
Q t + 1 i ( s t i , a t i ) =   Q t i ( s t i , a t i ; θ i ) + η ( R t + 1 +   γ m a x Q t i ( s t + 1 , a ; θ i ) Q t i ( s t i , a t i ; θ i ) ) +   j N a d j Q t 1 i ( s t 1 j , a t 1 j ; θ j )
where N a d j   is the number of adjacent intersections and θ j is parameter of evaluation network of intersection j.

5. Experiment

5.1. Experimental Setting

To evaluate the performance of cooperative traffic signal control with traffic flow prediction (TFP-CTSC) in a multi-intersection, the experiments are conducted on 4 × 4 road network scenarios as shown in Figure 5. Figure 5 shows our experiment road network and 2 × 2 intersection example which is part of a 4 × 4 intersection. TFP-CTSC is experimented in 16-intersection environment and each intersection is modeled as reinforcement learning agent. For cooperative traffic signal control, the number of vehicles in adjacent intersection are considered by transferring the state. We use Simulation of Urban Mobility (SUMO) [44] which is widely used in traffic signal control research. To receive the traffic status information, we use Traci API [45] supported in SUMO. TFP-CTSC is implemented using Pytorch which is python deep learning library. The hyperparameters for TFP-CTSC and configurations of road scenario and SUMO are shown in Table 2 and Table 3, respectively. In Table 2, from Batch size to Epoch are hyperparameters used for existing DQN, Alpha to prioritized epsilon for prioritized experience replay, Vmin to Atom size for distributional RL, and Multi-step size for multi-step learning. Table 3 shows the configurations of the road simulation environment built with SUMO.

5.2. Experimental Methodology

TFP-CTSC is trained to maximize the rewards. We define waiting time as a reward described before. In this experiment, we experimented with queue length and delay in addition to the waiting time used as a reward function to prove how effective the proposed algorithm is for improving overall traffic performance. The overall experimental structure and procedure is shown in Figure 6. The SUMO simulator creates a road environment using configuration files. In the road environment, vehicles are generated and observed periodically, and the observed vehicles are delivered to agents which were assigned to each intersection. Each agent selects the best traffic signal phase and switches the phase based on observed traffic volume. Through road conditions after traffic signals are switched, parameters such as the number of vehicles, queue length, waiting time, and delay are observed. The lower the waiting time, queue length, and delay, the better the performance. In the first experiment, we compare TFP-CTSC with existing reinforcement learning algorithms which are DQN, Q-learning. Through the first experiment, how well we define the reward function and our algorithm shows high performance in each three performance metrics. In the second experiment, we evaluate our main contributions which are cooperative method in multi-intersection and traffic flow prediction. We compare TFP-CTSC with the model, ignoring the traffic flow prediction and cooperative methods. In the final experiment, we evaluate the model performance according to the definition of the reward function most important in reinforcement learning.

5.3. Experimental Results

5.3.1. Comparative Analysis with Existing Algorithms

To evaluate TFP-CTSC, we compare our model with an existing reinforcement learning algorithm including DQN and Q-learning algorithm. The lower waiting time, queue and delay, the better the performance. Figure 7, Figure 8 and Figure 9 shows the overall performance of traffic signal control in a 4 × 4 intersection scenario for different performance metrics. It can be seen that TFP-CTSC model outperforms the other model. In three experimental results, we can see waiting time, queue length, and delay of TFP-CTSC is lower than other algorithm. And TFP-CTSC does not have much fluctuation compared to other models. That is, as the learning progresses, TFP-CTSC is more stable in both results of waiting time and queue.

5.3.2. Comparative Analysis for the Effect of TFP and CTSC

Our main contributions are cooperative traffic signal control (CTSC) and traffic flow prediction (TFP). To evaluate the effectiveness of TFP-CTSC, we compare TFP-CTSC with a model without CTSC and without CTSC and TFP. Figure 10a shows the result of comparing the average waiting time between each model and Figure 10b shows the result of queue length. In both Figure 10a,b, the TFP-CTSC outperformed the other models. Overall, TFP-CTSC shows lower waiting time and queue length, and tends to decrease as the training progresses.

5.3.3. Comparative Analysis of Reward Function

In this experiment, we compare the performance of the model according to how we define the reward function. The object of comparison is the reward function which considers only waiting time and the reward function, which considers both waiting time and queue length. Figure 11 shows the average reward at each intersection of two reward functions. The reward function considering only waiting time showed a high average reward at all intersections. This means that the reward function we defined has high performance and how to design the reward function is important.

6. Conclusions

In this paper, we propose a cooperative traffic signal control combined with traffic flow prediction (TFP-CTSC) at multiple intersections. Our traffic signal control algorithm has various extensions of DQN (Deep Q-Network) for fast learning speed and stable learning. We also introduce a method for cooperative traffic signal control to address the limitation of the lack of traffic signal control studies in a multi-intersection. Each intersection is modeled as one agent for cooperative traffic signal control, and each agent estimates the local Q-value. By transferring local Q-values between intersections, we can estimate global optimal Q-values. In addition, in order to consider variables that affect the actual traffic status, a traffic flow prediction model was separately constructed. The future traffic flow predicted by traffic flow prediction supports finding an optimal policy that maximizes the reward of the reinforcement learning model. To evaluate our model, we conducted various parametric experiments with existing reinforcement learning algorithms and demonstrated high performance. Also, to verify the effectiveness of our main contributions, cooperative method and traffic flow prediction, the proposed model is compared with the model without cooperative method and traffic flow prediction. As a result, this paper proposes cooperative traffic signal control combined with traffic flow prediction to mitigate traffic congestion at multiple intersections. In future work, we will study algorithms that cover wider road networks such as the city and improve them for practical applications.

Author Contributions

Conceptualization and Investigation, D.K.; Writing—Original Draft and Writing—Review & Editing, D.K.; Validation, D.K. and O.J.; Supervision and Project administration, O.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the NRF (National Research Foundation of Korea), and the MSIT (Ministry of Science and ICT), Korea, under the National Program for Excellence in SW supervised by the IITP (Institute for Information & communications Technology Promotion) (Nos. NRF 2019R1A2C1008412, 2015-0-00932).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aslani, M.; Stefan, S.; Marco, W. Continuous residual reinforcement learning for traffic signal control optimization. Can. J. Civ. Eng. 2018, 45, 690–702. [Google Scholar] [CrossRef]
  2. Lafferriere, G. A Decentralized Network Consensus Control Approach for Urban Traffic Signal Optimization; PDXScholar: Portland, OR, USA, 2019. [Google Scholar]
  3. Aslani, M.; Mesgari, M.S.; Seipel, S.; Wiering, M. Developing adaptive traffic signal control by actor–critic and direct exploration methods. Proc. Inst. Civ. Eng.-Transp. 2019, 172, 289–298. [Google Scholar] [CrossRef]
  4. Lv, Y.; Duan, Y.; Kang, W.; Wang, F.-Y. Traffic flow prediction with big data: A deep learning approach. In Proceedings of the IEEE Transactions on Intelligent Transportation Systems, Washington, DC, USA, 15–18 September 2015; pp. 865–873. [Google Scholar]
  5. Huang, W.; Song, G.; Hong, H.; Xie, K. Deep architecture for traffic flow prediction: Deep belief networks with multitask learning. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2191–2201. [Google Scholar] [CrossRef]
  6. Polson, N.G.; Vadim, O.S. Deep learning for short-term traffic flow prediction. Transp. Res. Part C Emerg. Technol. 2017, 79, 1–17. [Google Scholar] [CrossRef] [Green Version]
  7. Kühnel, N.; Theresa, T.; Kai, N. Implementing an adaptive traffic signal control algorithm in an agent-based transport simulation. Procedia Comput. Sci. 2018, 130, 894–899. [Google Scholar] [CrossRef]
  8. Williams, B.M.; Priya, K.D.; Donald, E.B. Urban freeway traffic flow prediction: Application of seasonal autoregressive integrated moving average and exponential smoothing models. Transp. Res. Rec. 1998, 1644, 132–141. [Google Scholar] [CrossRef]
  9. Pandit, K.; Ghosal, D.; Zhang, M.; Chuah, C.-N. Adaptive traffic signal control with vehicular ad hoc networks. IEEE Trans. Veh. Technol. 2013, 62, 1459–1471. [Google Scholar] [CrossRef] [Green Version]
  10. Mckenney, D.; White, T. Distributed and adaptive traffic signal control within a realistic traffic simulation. Eng. Appl. Artif. Intell. 2013, 26, 574–583. [Google Scholar] [CrossRef]
  11. Yang, K.; Tan, I.; Menendez, M. A reinforcement learning based traffic signal control algorithm in a connected vehicle environment. In Proceedings of the 17th Swiss Transport Research Conference (STRC 2017), Ascona, Switzerland, 17–19 May 2017. [Google Scholar]
  12. Bingham, E. Reinforcement learning in neurofuzzy traffic signal control. Eur. J. Oper. Res. 2011, 131, 232–241. [Google Scholar] [CrossRef]
  13. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, L.; Wierstra, D.; Riedmiller, M. Playing Atari with Deep Reinforcement Learning. 2013. Available online: https://arxiv.org/abs/1312.5602 (accessed on 19 December 2013).
  14. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  15. Zeng, J.; Hu, J.; Zhang, Y. Adaptive traffic signal control with deep recurrent Q-learning. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018. [Google Scholar]
  16. Wei, H.; Zheng, G.; Yao, H.; Li, Z. Intellilight: A reinforcement learning approach for intelligent traffic light control. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 19–23 August 2018. [Google Scholar]
  17. Wang, X.; Ke, L.; Qiao, Z.; Chai, X. Large-scale traffic signal control using a novel multi-agent reinforcement learning. arXiv Prepr. 2019, arXiv:1908.03761. [Google Scholar]
  18. Yang, S.; Yang, B.; Wong, H.-S.; Kang, Z. Cooperative traffic signal control using multi-step return and off-policy asynchronous advantage actor-critic graph algorithm. Knowl.-Based Syst. 2019, 183, 104855. [Google Scholar] [CrossRef]
  19. Mahmood, T.; Ali, M.E.M.; Durdu, A. A two stage fuzzy logic adaptive traffic signal control for an isolated intersection based on real data using SUMO simulator. Electron. Commun. Eng. 2019, 3, 656–659. [Google Scholar] [CrossRef]
  20. Lee, J.; Ganghwa, K.; Kitae, J. Estimation of road surface condition during summer season using machine learning. J. Korea Inst. Intell. Transp. Syst. 2018, 17, 121–132. [Google Scholar]
  21. Koesdwiady, A.; Ridha, S.; Fakhreddine, K. Improving traffic flow prediction with weather information in connected cars: A deep learning approach. IEEE Trans. Veh. Technol. 2016, 65, 9508–9517. [Google Scholar] [CrossRef]
  22. Sun, D.; Benekohal, R.F.; Waller, S.T. Bi-level programming formulation and heuristic solution approach for dynamic traffic signal optimization. Comput.-Aided Civ. Infrastruct. Eng. 2006, 21, 321–333. [Google Scholar] [CrossRef]
  23. Razzaq, M.; Shin, S. Fuzzy-logic dijkstra-based energy-efficient algorithm for data transmission in WSNs. Sensors 2019, 19, 1040. [Google Scholar] [CrossRef] [Green Version]
  24. Cheng, L.; Li, Y.; Wang, Y.; Bi, Y.; Feng, L.; Xue, M. A triple-filter NLOS localization algorithm based on fuzzy c-means for wireless sensor networks. Sensors 2019, 19, 1215. [Google Scholar] [CrossRef] [Green Version]
  25. Bi, Y.; Lu, X.; Sun, Z. Optimal type-2 fuzzy system for arterial traffic signal control. IEEE Trans. Intell. Transp. Syst. 2017, 19, 3009–3027. [Google Scholar] [CrossRef]
  26. Shiri, M.J.S.; Maleki, H.R. Maximum green time settings for traffic-actuated signal control at isolated intersections using fuzzy logic. Int. J. Fuzzy Syst. 2017, 19, 247–256. [Google Scholar] [CrossRef]
  27. Yang, W.; Zhang, L.; He, Z.; Zhuang, L. Optimized two-stage fuzzy control for urban traffic signals at isolated intersection and paramics simulation. In Proceedings of the 2012 15th International IEEE Conference on Intelligent Transportation Systems, Anchorage, AK, USA, 16–19 September 2012. [Google Scholar]
  28. Bookseller, A.A.; Jagtap, R.R. Image processing based adaptive traffic control system. IOSR J. Electron. Commun. Eng. 2013, 33–37. [Google Scholar]
  29. Ozan, C.; Baskan, O.; Haldenbilen, S.; Ceylan, H. A modified reinforcement learning algorithm for solving coordinated signalized networks. Transp. Res. Part C Emerg. Technol. 2015, 54, 40–55. [Google Scholar] [CrossRef]
  30. Wan, C.-H.; Hwang, M.-C. Value-based deep reinforcement learning for adaptive isolated intersection signal control. IET Intell. Transp. Syst. 2018, 12, 1005–1010. [Google Scholar] [CrossRef]
  31. Coşkun, M.; Baggag, A.; Chawla, S. Deep reinforcement learning for traffic light optimization. In Proceedings of the 2018 IEEE International Conference on Data Mining Workshops (ICDMW), Singapore, 17–20 November 2018. [Google Scholar]
  32. Li, L.; Lv, Y.; Wang, F.-Y. Traffic signal timing via deep reinforcement learning. IEEE/CAA J. Autom. Sin. 2016, 3, 247–254. [Google Scholar]
  33. Genders, W.; Razavi, S. Using a deep reinforcement learning agent for traffic signal control. arXiv Prepr. 2016, arXiv:1611.01142. [Google Scholar]
  34. Hasselt, H.V.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. In Proceedings of the Thirtieth AAAI conference on artificial intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  35. Wang, Z.; Schaul, T.; Hessel, M.; Hasselt, H.; Lanctot, M.; Freitas, N. Dueling network architectures for deep reinforcement learning. arXiv Prepr. 2016, arXiv:1511.06581. [Google Scholar]
  36. Hessel, M.; Modayil, J.; van Hasselt, H.; Schaul, T.; Ostrovski, G.; Dabney, W.; Horgan, D.; Piot, B.; Azar, M.; Silver, D. Rainbow: Combining improvements in deep reinforcement learning. In Proceeding of the Thirty-Second AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 2–7 February 2017. [Google Scholar]
  37. Ge, H.; Song, Y.; Wu, C.; Ren, J.; Tan, G. Cooperative deep q-learning with q-value transfer for multi-intersection signal control. IEEE Access 2019, 7, 40797–40809. [Google Scholar] [CrossRef]
  38. Schaul, T.; Quan, J.; Antonoglou, I.; Silver, D. Prioritized experience replay. arXiv Prepr. 2016, arXiv:1511.05952. [Google Scholar]
  39. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  40. Bellemare, M.G.; Dabney, W.; Munos, R. A distributional perspective on reinforcement learning. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  41. Fortunato, M.; Azar, M.G.; Piot, B.; Menick, J.; Osband, I.; Graves, A.; Mnih, V.; Munos, R.; Hassabis, D.; Pietquin, O.; et al. Noisy networks for exploration. arXiv Prepr. 2017, arXiv:1706.10295. [Google Scholar]
  42. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  43. Witwicki, S.; Castillo, J.C.; Messias, J.; Capitan, J.; Melo, F.S.; Lima, P.U.; Veloso, M. Autonomous surveillance robots: A decision-making framework for networked muiltiagent systems. IEEE Robot. Autom. Mag. 2017, 24, 52–64. [Google Scholar] [CrossRef]
  44. Krajzewicz, D.; Erdmann, J.; Behrisch, M.; Bieker, L. Recent development and applications of SUMO-simulation of urban mobility. Int. J. Adv. Syst. Meas. 2012, 5, 128–138. [Google Scholar]
  45. Wegener, A.; Piórkowski, M.; Raya, M.; Hellbrück, H.; Fischer, S.; Hubaux, J.-P. TraCI: An interface for coupling road traffic and network simulators. In Proceedings of the 11th Communications and Networking Simulation Symposium, Ottawa, ON, Canada, 14–17 April 2008. [Google Scholar]
Figure 1. Reinforcement learning based traffic signal control.
Figure 1. Reinforcement learning based traffic signal control.
Sensors 20 00137 g001
Figure 2. The 4-arm structure of intersection.
Figure 2. The 4-arm structure of intersection.
Sensors 20 00137 g002
Figure 3. Overall architecture of TFP-CTSC.
Figure 3. Overall architecture of TFP-CTSC.
Sensors 20 00137 g003
Figure 4. (a) Correlation matrix and (b) Visualized results of traffic flow prediction.
Figure 4. (a) Correlation matrix and (b) Visualized results of traffic flow prediction.
Sensors 20 00137 g004
Figure 5. 4 × 4 intersection road network.
Figure 5. 4 × 4 intersection road network.
Sensors 20 00137 g005
Figure 6. Experimental structure and procedure.
Figure 6. Experimental structure and procedure.
Sensors 20 00137 g006
Figure 7. Performance comparison for average waiting time.
Figure 7. Performance comparison for average waiting time.
Sensors 20 00137 g007
Figure 8. Performance comparison for queue length.
Figure 8. Performance comparison for queue length.
Sensors 20 00137 g008
Figure 9. Performance comparison for delay.
Figure 9. Performance comparison for delay.
Sensors 20 00137 g009
Figure 10. Performance comparison with TFP-CTSC without CTSC and TFP. (a) Average waiting time; (b) queue length.
Figure 10. Performance comparison with TFP-CTSC without CTSC and TFP. (a) Average waiting time; (b) queue length.
Sensors 20 00137 g010
Figure 11. Performance comparison for reward function.
Figure 11. Performance comparison for reward function.
Sensors 20 00137 g011
Table 1. The example of traffic signal system at single intersection.
Table 1. The example of traffic signal system at single intersection.
Traffic Signal ID = ‘2’
DirNESW
Phase RSLRSLRSLRSL
Phase 0 (NSG)GGrr r rGGrr r r
Phase 1y y y r r ry y y r r r
Phase 2 (NSLG)yyGr r ryyGr r r
Phase 3rryr r rrryr r r
Phase 4 (EWG)r r rGGrr r rGGr
Phase 5r r ryy rr r ryy r
Phase 6 (EWLG)r r rrrGr r rrrG
Phase 7r r rryrr r rryr
Table 2. Hyperparameters for TFP-CTSC.
Table 2. Hyperparameters for TFP-CTSC.
HyperparameterDescriptionValue
Batch sizeSize of batch for sampling32
Learning rate Step size in loss function0.0001
Discount factorWeight that multiplies future reward0.99
Replay memory sizeSize of Replay memory1000
OptimizerOptimization algorithmAdam
Target network replacement frequencyHow frequently update target network100
EpochTraining step size2000
AlphaHow much prioritization used0.2
BetaHow much importance sampling is used0.6
Prioritized epsilonHow much every transition can be sampled1e-6
VminMinimum value of support -10
VmaxMaximum value of support10
Atom sizeThe unit number of support51
Multi-step sizeStep size to calculate n-step reward3
Table 3. Configuration of Road scenario and SUMO.
Table 3. Configuration of Road scenario and SUMO.
HyperparameterValue
Intersection size16
Number of lane2
Simulation start 0 s
Simulation end 35,999 s
Vehicle speed13.00 km/h
Duration of yellow2.0 s
Minimum duration of green3.0 s
Maximum duration of green50.0 s

Share and Cite

MDPI and ACS Style

Kim, D.; Jeong, O. Cooperative Traffic Signal Control with Traffic Flow Prediction in Multi-Intersection. Sensors 2020, 20, 137. https://doi.org/10.3390/s20010137

AMA Style

Kim D, Jeong O. Cooperative Traffic Signal Control with Traffic Flow Prediction in Multi-Intersection. Sensors. 2020; 20(1):137. https://doi.org/10.3390/s20010137

Chicago/Turabian Style

Kim, Daeho, and Okran Jeong. 2020. "Cooperative Traffic Signal Control with Traffic Flow Prediction in Multi-Intersection" Sensors 20, no. 1: 137. https://doi.org/10.3390/s20010137

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop