Next Article in Journal
Prediction of Crack Growth Life at Elevated Temperatures with Neural Network-Based Learning Schemes
Next Article in Special Issue
A Policy-Reuse Algorithm Based on Destination Position Prediction for Aircraft Guidance Using Deep Reinforcement Learning
Previous Article in Journal
Optimization Framework for Early Conceptual Design of Helicopters
Previous Article in Special Issue
Dynamic Prediction of Air Traffic Situation in Large-Scale Airspace
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image-Based Multi-Agent Reinforcement Learning for Demand–Capacity Balancing

Department of Computer Architecture, Escola d’Enginyeria de Telecomunicació i Aeroespacial de Castelldefels (EETAC), Universitat Politècnica de Catalunya (UPC), Esteve Terradas 7, 08860 Castelldefels, Barcelona, Spain
*
Author to whom correspondence should be addressed.
Aerospace 2022, 9(10), 599; https://doi.org/10.3390/aerospace9100599
Submission received: 26 August 2022 / Revised: 7 October 2022 / Accepted: 12 October 2022 / Published: 14 October 2022
(This article belongs to the Special Issue Advances in Air Traffic and Airspace Control and Management)

Abstract

:
Air traffic flow management (ATFM) is of crucial importance to the European Air Traffic Control System due to two factors: first, the impact of ATFM, including safety implications on ATC operations; second, the possible consequences of ATFM measures on both airports and airlines operations. Thus, the central flow management unit continually seeks to improve traffic flow management to reduce delays and congestion. In this work, we investigated the use of reinforcement learning (RL) methods to compute policies to solve demand–capacity imbalances (a.k.a. congestion) during the pre-tactical phase. To address cases where the expected demands exceed the airspace sector capacity, we considered agents representing flights who have to decide on ground delays jointly. To overcome scalability issues, we propose using raw pixel images as input, which can represent an arbitrary number of agents without changing the system’s architecture. This article compares deep Q-learning and deep deterministic policy gradient algorithms with different configurations. Experimental results, using real-world data for training and validation, confirm the effectiveness of our approach to resolving demand–capacity balancing problems, showing the robustness of the RL approach presented in this article.

1. Introduction

Congestion problems arise in situations where limited resources have to be shared simultaneously by multiple agents. They are present in a wide variety of domains in the modern world, and they have drawn much attention in Artificial Intelligence (AI) research [1,2]. Air Traffic Management (ATM) is one domain where congestion problems appear naturally, introducing extra costs and uncertainty to operations scheduling. Concretely, congestion problems appear when the expected number of aircraft (airspace demand) exceeds the maximum number of flights that the Air Traffic Controllers (ATCOs) can safely manage for a particular airspace sector (capacity). This is known as the Demand–Capacity Balancing (DCB) problem or process [3].
Initially, demand–capacity imbalances are solved via airspace management or flow management solutions. However, when no solution is enough, Air Traffic Flow Management (ATFM) regulations are implemented issuing extra ground delays to the necessary flights. This cascade of events increases the uncertainty regarding the scheduling of operations, costs [4], and unforeseen effects on the entire system. Furthermore, these events present further negative effects for the ATM stakeholders, including environmental effects, customer satisfaction, and loss of reliability.
In the European Air Traffic Control (ATC) network, ATFM delays are imposed by the Computer Assisted Slot Allocation (CASA) algorithm [3,5], which is a heuristic algorithm based on the principle of first-planned-first-served. In 2018, prior to COVID-19, at the European Civil Aviation Conference (ECAC) level, the number of flights increased by +14.6%, corresponding to 1.4 million additional flights in 2018 compared to 2013. At the same time, en-route ATFM delays more than doubled compared to 2017 (+104%). As a result, 9.6% of the flights issued ATFM delays with an increment of 1.74 min per flight [6].
Nowadays, demand–capacity imbalances are difficult to predict during the pre-tactical phase (from several days to a few hours prior to operations) because of the uncertainty in the operational information. Indeed, DCB is a two-stage problem: first, it is necessary the identification of the overloaded regions (demand greater than capacity); second, the CASA algorithm assigns new departure slots smoothing the demand to meet the pre-defined capacity.
Previous research on ATFM regulations focused on the detection and/or resolution of DCB issues, optimization algorithms, development of new performance metrics, or novel techniques. For instance, in previous research [7,8], we proved the potential of supervised Machine Learning (ML) models to predict which sectors were going to be regulated, Reference [9] used ML models to predict the evolution of the ATFM delay for regulated flights, and Reference [10] investigated the detection of regulations due to convective weather and the associated airspace performance characteristics.
Other works investigated optimization techniques to find optimal resource utilization. Reference [11] presented an optimization algorithm to minimize the propagation of ATFM delays to subsequent flights, Reference [12] introduced an integer programming model for strategic redistribution of flights to respect nominal sector capacities in short computation times for large scales and Reference [13] investigated a new technique that could improve airspace capacity usage and reduce ATFM delays by improving the slot allocation process of CASA to avoid wasted capacity (empty slots) in regulated sectors.
On the other hand, several works attempted to study the downstream effects of ATFM regulations and propose resolution techniques. Reference [14] used gradient-boosted decision trees to predict the likelihood of a regulated flight rerouting to mitigate the ATFM delay, and Reference [15] proposed using speed reduction on air to absorb ATFM delay at no extra cost. Most recent works on the resolution of DCB issues focused on the use of Reinforcement Learning (RL) techniques. For instance, Reference [16] proved it was possible to both identify and resolve DCB problems by comparing three RL algorithms for the pre-tactical phase. Similarly, Reference [17] is the result of a set of publications where the DCB was formulated as a hierarchical Multi-Agent Reinforcement Learning (MARL) decision problem with different levels of abstraction. However, one important drawback of this MARLs approach is that a different agent controls each flight, presenting a severe scalability problem, as hundreds or even thousands of different agents would be required to handle the full European Air Traffic Management Network (EATMN).
In response to the previous scalability limitations, Reference [18] presented a collaborative Multi-Agent Asynchronous Advantage Actor-Critic (MAA3C) framework with embedded supervised and unsupervised Neural Network (NN), where only flights crossing airspace sectors with already identified demand–capacity issues were regarded as the candidate agents. This approach improved the scalability and generalization of the system, being able to handle a varying number of agents. As an extension of the scalability issues, Reference [19] reviewed different deep MARL methods, examining their ability to scale up to large agent populations (from hundreds to several thousands of agents). The main conclusion drawn with respect to possible scalability issues is the importance of parameter sharing in large agent populations. It is impractical to train thousands of independent networks for each agent or to utilize an approach whose input size would explode as the number of agents and their observations grow larger.
Similar research has been conducted outside the EATMN. In the USA network, Reference [20] developed a MARL system for ATM integrated with an air traffic flow simulator—FACET. In Reference [21], the authors presented a distributed decision support system for tactical ATFM in Brazil and traffic flow managers experts analyzed the solutions proposed by the system.
Traffic growth and changes in traffic patterns have caused increasing congestion and delay in European airspace. The Central Flow Management Unit (CFMU) continually seeks and develops methods to improve traffic flow management to reduce delays and congestion [5]. To this end, and taking into account the available literature, as a research question, this article investigates whether a RL system is able to solve demand–capacity imbalances for specific airspace sectors without sharing explicit information between agents, using an approach whose size does not depend on the number of agents and with a fixed size of the observation states.
This article formalizes the ATFM problem as a collaborative MARL system where homogeneous agents, representing flights, aim to decide on their ground delay jointly with the other flights, while not having direct information about the preferences of other flights. The goal is to smooth already-identified pre-tactical DCB problems in a specific airspace sector using images as input to the system and ensuring efficient utilization of the airspace. The usage of images allows the system to extract its own features for the problem instead of manually deciding which ones are more representative, the input size is independent of the number of agents, and it provides a fixed size of the states ensuring good scalability. Moreover, the images allow the agents to have indirect information about the other flights.
To smooth the demand–capacity imbalances, we investigated two types of RL algorithms: first, algorithms based on discrete actions; second, algorithms based on continuous actions. In both cases, we used a homogeneous population of agents to ensure their behavior was the same. Furthermore, the agent-based paradigm introduced in this article tried to emulate the first-planned-first-served basis used in the current ATFM approach. Only flights outside the regulated sector were candidates to be agents, ensuring that only flights outside the airspace sector would be delayed.
Experimental results demonstrated the effectiveness and robustness of the novel approach presented in this paper. The work presented in this article aimed to be the first step toward devising multi-agent methods for deciding on delay policies using images. The contributions made in the paper are as follows:
  • The DCB problem was formulated as a MARL based on the use of images to improve scalability.
  • Two different types of MARL algorithms were studied, one using continuous actions (DQN) and another one using discrete actions (DDPG).
  • For the DDPG algorithm, three different approaches for exploration noise (random values used for trial-and-error search) were analyzed.
  • All configurations were trained and evaluated using real-world data.
This paper is organized as follows: Section 2 provides relevant background about RL, focusing on Q-learning and deterministic policy gradient. Section 3 presents the problem formulation. Section 4 describes the experimental setup focusing on the parameters of the algorithms. Section 5 exhibits the experimental results obtained and their analysis. Section 6 summarizes the work presented in this article and addresses future work.

2. Reinforcement Learning

RL problems consist of learning what to do (how to map situations to actions) to maximize a numerical reward signal. The agent is not told which actions to take, but it must discover which actions yield the most reward by trying them. Notice that actions may affect not only the immediate reward but also the following states and, through that, all subsequent rewards. These two characteristics, trial-and-error search, and delayed reward, are the two most important distinguishing features of RL. Therefore, a learning agent must be able to sense the state of the environment, take actions that affect the state, and have a clear goal (or goals) relating to the state of the environment [22]. This interaction is depicted in Figure 1.
One of the challenges that arise in RL is the trade-off between exploration and exploitation. To obtain as much of a reward as possible, a RL agent must prefer actions that it tried in the past and found to be effective. However, to discover them, it has to try actions it has not selected previously. The agent has to exploit what it has already experienced to maximize the reward, but it has to explore to make better action selections in the future.
Beyond the environment and the agent, we can identify four main sub-elements:
  • Policy: roughly speaking, it is a mapping from the states of the environment to actions.
  • Reward signal: it defines the goal of the RL problem. It defines what is “good” in an immediate sense.
  • Value function: specifies what is “good” in the long run. Roughly speaking, it is the total reward an agent can expect to accumulate over the future, starting from a particular state.
  • Model: it mimics the behavior of the environment. It allows inference to be made about how the environment will behave.

2.1. Single-Agent Reinforcement Learning

A RL problem for a single agent interacting with an environment can be formalized as a finite Markov Decision Process (MDP) described by the tuple ( S , A , P , R ), where S is the set of states of the environment, A is the set of actions the agent can take, P is the transition function, being P ( s | s , a ) the probability of transitioning to s S , by applying a A in s S , and R is the reward function. Notice, in a finite MDP, the sets of states, actions, and rewards ( S , A , and R) have a finite number of elements.
At each time step, the reward is a simple number, R t R . However, the agent’s goal is to maximize its cumulative reward G. That is, maximize both the immediate reward and cumulative reward in the long run. Thus, the rewards we set up must truly indicate what we want to accomplish. The cumulative reward, also referred to as return, can be defined as follows:
G = t = 0 γ t R t
where γ is a parameter, 0 < γ < 1 , called the discount rate. It determines how much the agent cares about immediate rewards relative to distant ones.
The RL system aims to find the optimal policy π * , which maximizes the expected commutative reward. Let us define the value function of a state v π ( s ) , for a policy π (which may not be optimal), as the expected return when starting in state s following policy π . For MDPs, we can define v π ( s ) , formally by:
v π ( s ) = E [ G t | S t = s ]
Similarly, we define the value of taking action a in state s under a policy π as q π ( s , a ) , providing the expected return:
q π ( s ) = E [ G t | S t = s , A t = a ]
There is always at least one policy that is better than or equal to all other policies. The optimal policy. Although there may be more than one, all the optimal policies are denoted by π * . They share the same state-value function, called the optimal state-value function, denoted v * and defined as:
v * ( s ) = max π v π ( s )
for all s S .
Optimal policies also share the same optimal action-value function, denoted q * and defined as
q * ( s ) = max π q π ( s , a )
for all s S and a A .
Therefore, the optimal policy π * selects what action maximizes the expected commutative reward. If the optimal action-value function q * ( s , a ) is known, the best action in the state s is given by:
π * ( s ) = a r g max a A q * ( s , a )
The two main approaches used to obtain the optimal policy are policy iteration, which manipulates the policy directly, and value iteration, which aims to find an optimal value function adopting a greedy policy.

2.2. Multi-Agent Reinforcement Learning

A MARL system involves a set of N interacting agents, which can be cooperative, competitive, or both. It can be described by the tuple ( N , S , { A i } i N , { O i } i N , P , { R i } i N ) .
At every time step, each agent i N observes a partial representation of the environment o i O i , and performs an action a i A i determined by a policy function π i . Then, when an action is taken, the environment evolves to a new state s S , according to the transition function P. This transition function depends on the current state and the joint action of all agents. Finally, the reward that each agent receives is given by the reward function. For instance, agents typically share the reward in a cooperative RL.
One possible approach for MARL is to train independent agents. However, this simple approach does not perform well in practice [24]. To overcome these limitations, in [25,26], each agent has its centralized critic, only used during learning, which approximates and learns the action-value function given the observations and actions of all agents. However, the critics require the actions and observations of all agents as input. Consequently, their complexity is proportional to the number of agents.
A different solution is proposed in [27] to mitigate this scalability issue. In this case, the agents learn an individual action-value function based on their local observations, and the sum of these functions approximates the centralized joint action-value function.

2.3. Q-Learning

Q-learning [23] is one of the most well-known algorithms based on value iterations. It makes use of a Q-table, which typically has the shape [states, actions], and each Q-value Q ( s , a ) represent the quality of taking as action a A , in s S . Thus, Q-learning was designed to work with discrete actions.
At each time step Δ t , the agent observes the current state s t and chooses the action a t with the highest Q-value in that state. After applying the selected action, the agent receives a reward r t , enters on new state s t + 1 , and the Q-value is updated using Equation (7):
Q ( s t , a t ) Q ( s t , a t ) + α r t + γ max a Q ( s t + 1 , a t ) Q ( s t , a t )
where r t is the reward received when moving from state s t to s t + 1 , α [ 0 , 1 ] is the learning rate and γ [ 0 , 1 ] is the discount factor.
According to Equation (7), the agent adopts a greedy strategy by constantly selecting the actions with the largest Q-value. In that case, it exists the risk of adopting a sub-optimal solution by converging to a local minimum. The ϵ -greedy strategy is widely used to properly explore the state space, where ϵ corresponds to the probability of choosing a random action. Typically, ϵ is usually initialized to 1 to force high exploration at the beginning, with a decay rate over time to ensure exploitation at the end of the training.
One limitation of this well-known algorithm is the rapid growth of dimensionality in the state space. The traditional solution is deep Q-learning [28], which uses a NN to approximate the Q-values. However, instead of training the NN with the sequence of experiences as they occur during the simulations, they are saved in what is usually called the experience replay buffer. Using a buffer prevents the agent from forgetting past experiences as time evolves and breaks the correlation between consecutive experiences. Finally, a target network is used to stabilize the learning. The target network is the result of periodically replacing its weights with the ones from the online network used to select the action greedily.

2.4. Deterministic Policy Gradient

Deterministic Policy Gradient (DPG) [29] is an actor–critic RL algorithm, used for continuous actions, which learns a deterministic policy function and a value function simultaneously, from an exploratory behavior.
It is not possible to straightforwardly apply Q-learning to continuous action spaces because finding the greedy policy would require optimization of a t at every time step, which is too slow to be practical with large, unconstrained functions approximators, and nontrivial action spaces [29]. The DPG algorithm uses an actor as the current policy to map states to a specific action. The critic determines the expected reward for an agent starting at a given state and acting according to the previous policy.
As with Q-learning, to learn and generalize on large-scale state spaces, it is required to introduce non-linear function approximators, which means that convergence is no longer guaranteed. However, such approximators appear essential in those scenarios [30] that presented a modification to DPG from [31], inspired by the success of Deep Q-Learning (DQN), allowing the use of NN function approximators. This implementation is called Deep Deterministic Policy Gradient (DDPG), and it was proved that the algorithm could learn policies “end-to-end” directly from raw pixel inputs. Target networks are used to add stability to the training, and an experience replay buffer is used to learn from experiences accumulated during the training.

3. Problem Formulation

In the current European ATC network, ATFM delays are particularly complex problems. When flights are affected by an ATFM delay, they are issued with a Calculated Take-Off Time (CTOT), which indicates the new time windows for the flight to depart (from 5 min before the CTOT to 10 min after). This new CTOT is computed by the CASA algorithm, and if the flights cannot depart within this window, the ATFM slot will be missed and a new one will be assigned.
ATFM regulations are located at specific airspace traffic volumes (which can be informally defined as a portion of airspace linked to a sector) where a demand–capacity imbalance is detected. Nowadays, the methodology used to identify where ATFM regulations are required is purely human and does not rely on automation. Air Navigation Service Providers (ANSPs) define two capacities for the sectors, which have to be interpreted by the Flow Manager Position (FMP): the sustained capacity and the peak capacity. The sustained capacity indicates the maximum number of flights that can be operated for a particular time window, while the peak capacity indicates the maximum value for a specific instant of time. Close to the day of operation, capacities are defined based on the Occupancy Count (OC), which considers the expected number of flights inside the traffic volume.
It is possible to have multiple demand–capacity imbalances in the network simultaneously. However, the general principle is that a flight subject to several ATFM regulations is given the delay of the most penalizing regulation.

3.1. Assumptions

In this work, the following assumptions are considered to define the ATFM delay (a.k.a. ground delay) system for specific traffic volumes:
  • The airspace sectors with a demand–capacity imbalance are known (interval of time with overload, location, and capacity), and squares can be used to approximate their shape.
  • Pre-tactical flight plans are available for each flight before any regulation is applied. The flight plans contain the Scheduled Off-Block Time (SOBT) and the route of the flight. Additionally, it is assumed constant speed for each of the segments composing the routes.
  • There is one type of agent. There are no aircraft with priority.
  • Financial costs imposed on commercial entities resulting from ATFM decisions are negligible.
There is a deviation from traditional state-of-the-art problems by assuming the demand–capacity imbalances are already known for the sector of study. Assumption 1 was used because this work focused purely on the resolution of DCB issues. Only historical data from regulated intervals and sectors were used. Moreover, related to assumption 1, the approximation of the sector’s shape as squares aimed to reduce the implementation complexity in this preliminary study.
Related to assumption 2, we considered the constant speed per segment defining the routes because they only contained information about the starting/ending location and time. By assuming a constant speed between the origin and end of the segments, it is possible to interpolate the location of the flights at intermediate timestamps (see [32,33] as other examples of interpolation).
Assumption 3 aspires to create a prototype that is as fair as possible for all the operators. A homogeneous population of agents guarantees that all flights are treated equally. However, using heterogeneous populations of agents in future work could be interesting from an optimization point of view. For instance, the use of different populations between domestic or international flights, or to prioritize transit flights to avoid possible downstream effects such as missing connections. Similarly, assumption 4 is used to emphasize that this prototype focuses on the current used Key Performance Indicators (KPIs), although they could be extended according to additional requirements if needed.

3.2. Action Variable

The action variable in this problem corresponds to selecting the ground delay that an aircraft will receive due to a demand–capacity imbalance. At each step Δ t , each agent i N has an associated action variable a t i A , where a t i is the ATFM ground delay.
For discrete action algorithms, the action variable can be defined as:
a t i A , A { 0 , 5 , 10 , 15 }
While for continuous action algorithms, the action variable can be defined as:
a t i A , A [ 0 , 15 ]

3.3. State Variable

The state vector s t i S includes the information that the population of agents N uses to determine the actions. Each state s t i is defined per flight candidate to be an agent and step of the system.
One of the primary challenges associated with MARL is problem representation. The challenge is in defining the problem in such a way that an arbitrary number of agents can be represented without changing the architecture of the DQN or DDPG. To solve this problem, we propose the usage of image-like tensors where each channel in the images encodes a different set of information from the global state. This representation allows us to take advantage of Convolutional Neural Networks (CNNs), which have been shown to work well for image classification tasks [34] and competitive MARL systems based on images [35].
The image tensor is of size HxWx3 (shown in Figure 2), where H is the height, W is the width of our two-dimensional images, and three is the number of channels in the image. The channels can be broken down in the following way:
  • Inside channel: Contains the representation of the flights inside the sector being regulated.
  • Outside channel: Contains information about the flights outside the sector of study, i.e., the flights that may be delayed.
  • Self channel: Contains information about the agent making the decision.
Note that the three channels are depicted with white backgrounds for clearness, but they encode zero pixel values. The non-zero pixel values encode the locations of the flights, their headings, and the approximate shape of the sector.

3.4. State Transition

The state transition defines a set of conditions that determine how the state s t i S evolves along the steps. With every step, the aircraft candidates (to be agents) decide whether they are going to issue ATFM delays. Three conditions must be verified to ensure a proper transition between states.
The first condition that must be verified is related to the regulations used for the training. Each episode will start using information from a randomly selected historical regulation, and the environment will evolve for a time period T P equal to 60 min with a timestep Δ t equal to one minute. Thus, from the randomly selected regulation, we must guarantee that the regulations will be active for more time than the T P .
The second condition to consider is related to the delay. For each state variable s t i S , the agent i will produce a new action to cooperatively decide its own ground delay to ensure that the demand meets the capacity. Actions equal to zero imply no delay for the flight moving forward on the predefined trajectory. However, if the delay differs from zero, the new delay is added to possible previous delays (cumulative delay).
The last required consideration is related to how the flight is assumed to move forward. A trajectory T T is a time series of segments of the form:
T = { ( I D l , b e g i n t l , e n d t l , l a t _ b e g i n t l , l o n _ b e g i n t l , l a t _ e n d t l , l o n _ e n d t l ) } l [ 1 , m ]
where I D l is the identifier of the segment, b e g i n t l the initial timestamp of the segment, e n d t l the end timestamp of the segment, l a t _ b e g i n t l , l o n _ b e g i n t l the initial latitude and longitude of the segment, l a t _ e n d t l , l o n _ e n d t l the end latitude and longitude of the segment, and l is the number of segments used to define the trajectory.
For each of the segments, we assumed constant speed. Therefore, the expected velocity of the flight in a particular segment can be defined as follows:
v I D l = f ( l a t _ e n d t l , l a t _ b e g i n t l , l o n _ e n d t l , l o n _ b e g i n t l e n d t l b e g i n t l
where f is a function that computes the distance between two pairs of coordinates.
Finally, we can compute the aircraft’s location at any timestamp, knowing the required segment to use, assuming constant speed in the segments, and taking into account the imposed ATFM delay.

3.5. Problem Constrain

A set of operational constraints is associated with the sectors’ capacity, whose violation results in DCB problems (congestion). This violation occurs when O C T P C , where O C T P is the demand of a predefined counting period and C is the capacity of the sector. Although this violation will result in a DCB issue, it must be taken into account that two capacities are defined: sustain and peak capacity. The sustained capacity can be exceeded for a short period, while the peak capacity should not be exceeded. Furthermore, for each new episode, it must be guaranteed that there is no demand–capacity imbalance for the selected random starting timestamp of the ATFM regulation. Ensuring that the agents are not directly penalized due to the demand–capacity imbalance without being able to perform any action.
Another constraint is related to which flights are considered candidates to be agents. Thus, the flights can be delayed in each step. To emulate as close as possible the current approach based on first-in-first-served used by CASA, for each step, only those flights outside the congested sector are considered agents.
The last constraint of the system is related to the ATFM delay itself. The maximum delay an agent can impose on a flight per step is equal to 15 min. This aims to allow agents to delay flights without imposing huge penalization.

3.6. Objective Function

Demand reduction is one of the main goals in DCB during the pre-tactical phase. The objective is to try to smooth the traffic and meet the expected demand with the predefined capacity of the airspace sector. The objective function can be defined with Equation (12), which corresponds to minimizing the ATFM delay while trying to ensure that the demand meets the sector’s capacity for the counting period.
min t T P E { i = 0 N D i s t , π * ( s t ) } V t C
where D i is the ATFM delay of agent i, N is the population of the agent, s t is the state of the system at step t, π * is the optimal ATFM delay policy, V t is the OC of the sector, and C is the capacity of the sector.

4. Experimental Setup

This section details the developed DQN and DDPG algorithms, focusing on the dataset used to train the agents, the RL elements, and the parameters of the algorithms.

4.1. Dataset

The proposed RL algorithms are trained using pre-tactical information from Aeronautical Information Regulation and Controls (AIRACs). Concretely, we used data from June, July, August, and September 2019.
In the EATMN, a wide variety of regulations are applied due to many reasons across different traffic volumes. The study done in this article focuses on C-ATC Capacity ATFM regulations, which are those regulations purely related to demand–capacity imbalances. Other types of regulations can be related to convective weather or military operations. Moreover, because of the huge number of sectors, we focused our attention on the Maastricht Upper Area Control Centre (MUAC) region. In particular, to the sector EDYYBOLN with the associated traffic volume MASBOLN. The main reason behind the selection of this particular sector is because it is one of the most regulated airspace regions in the MUAC area, which will guarantee enough variety of samples to train the RL agents. The available dataset contains around 200 C-ATC Capacity ATFM regulations for en-route traffic along 71 different days, with a mean number of regulations per day equal to 1.7 and a mean duration per regulation of 97.08 min.

4.2. Reward Function

RL algorithms learn from the interactions with an environment, which provides a reward according to how good the agent’s action was. The reward function is crucial because different reward structures will result in different system performances.
Previous research has investigated different reward functions. Typically, the literature shows that researchers mainly focused on delay and congestion without considering fairness impact on different commercial entities [20,36]. Similarly, [37] also took into account the amount of time the agents contributed to the demand–capacity imbalance. Fairness is usually measured by financial costs imposed on commercial entities resulting from ATFM decisions [38].
In our case, as a proof of concept using images, focus on delay and congestion. The reward function G ( z ) , written as Equation (13), consists of three main components: the number of flights delayed C ( z ) , the delay itself D ( z ) , and the demand–capacity ratio I ( z ) :
G ( z ) = β M ( z ) δ D ( z ) λ I ( z )
where z represents the system under evaluation, M ( z ) and D ( z ) represent delay, and I ( z ) the congestion; β , δ , and λ are the weights used to adjust the income penalty in the evaluation function. Note that the reward function is based on penalties.
When ATFM delay is issued, the number of aircraft entering the airspace traffic volume is reduced; thus, the congestion is relieved. However, this restrictive measure has negative effects on the ATM network. Equation (14) counts the number of delayed flights and Equation (15) computes the total delays imposed:
M ( z ) = Θ N
where N is the population of agents, and Θ is a function that counts the number of flights that received ATFM delay.
D ( z ) = t = 0 P d i N t
where D is the total ground delay, P is the counting period, and r i t is the imposed ground delay at step t for each agent i N .
It is required to compute the number of aircraft at the current step to determine the congestion severity in the airspace sector; that is, the excessive number of aircraft in the sector. The congestion function I ( z ) is given by:
I ( z ) = { ( V C ) ( V C ) V > C 0 O t h e r w i s e
where V is the number of aircraft in the sector, and C is the maximum number of aircraft in a sector that does not cause congestion (capacity). Note that the function is characterized exponentially with respect to the excessive number of aircraft in a sector.

4.3. Deep Q-Learning

In this work, the first RL algorithm we studied to optimize the ATFM delay was DQN. We followed the approach proposed in [39], which operates directly on RGB images to play Atari games, uses experience replay to store the agents’ experiences, and uses a second target network.
At the beginning of each episode, a new initial state is set. Subsequently, for each step and flight candidate to be an agent, an action is chosen either randomly or greedily and stored in the replay buffer. In the first episode, the ϵ -greedy strategy has an ϵ equal to 1 forcing agents to explore. However, this value linearly decreases until it reaches 0.01, ensuring the agents prioritize exploitation in the last episodes.
The input to the system is the images used to obtain the agent’s experience tuple of the form ( s t , a t , r t , s t + 1 ) , where s t is the starting image-like state, a t is the action taken, r t is the reward received, and s t + 1 is the new state of the system. The replay buffer stores the last 25,000 experience tuples, and batches with 64 samples are randomly selected to train the NN computing the target value and the respective loss. This loss is the minimum squared error of the predicted value and the target value, and the Adam optimizer [40] is used. After training the online network, the weights of the target network are also updated.
The input layer of the NN takes as input the 150 × 100 × 3 images. The first layer convolves 32 (8 × 8) filters with stride 4 and uses a Rectified Linear Unit (ReLU) activation function. The second layer is a batch normalization layer [41]. The third layer convolves 64 (4 × 4) filters with stride 2 using a ReLU activation function. The fourth layer is a batch normalization layer. The fifth layer convolves 64 (3 × 3) filters with stride 1 and uses a ReLU activation function. The sixth layer is a batch normalization layer. The final hidden layers are a fully-connected with 256 rectifier units and a droupout layer with a rate of 0.5. The output layer is a fully-connected linear layer with a single output for each valid action. The output of the NN corresponds to the predicted Q-values of the individual action for the input state. The main advantage of this type of architecture is the ability to compute Q-values for all possible actions in a given state with only a single forward pass through the network. Table 1 shows the remaining hyperparameters.

4.4. DDPG

The second algorithm we want to study to optimize ATFM delays is DDPG. We follow the approach presented in [30], which adapts the ideas underlying the success of DQN to continuous actions. DDPG is an actor–critic method, where a parameterized actor function μ ( s ) specifies the current policy by mapping states to actions while the critic Q ( s , a ) learns how good is the action. Similarly to our previous approach, this implementation of DDPG directly learns from raw pixel information, using a replay buffer, and throughout the use of target networks (one for the actor and one for the critic).
The chosen NN for the actor takes as input 150 × 100 × 3 images. The first layer convolves 32 (8 × 8) filters with stride 4 and uses a ReLU activation function. The second layer is a batch normalization layer. The third layer convolves 64 (4 × 4) filters with stride 2 and uses a ReLU activation function. The fourth layer is a batch normalization layer. The final hidden layers are a fully-connected with 256 rectifier units and a droupout layer with a rate of 0.5. The output layer is a fully-connected linear layer with a single output unit.
The chosen NN for the critic takes as input 150 × 100 × 3 images and the action predicted by the actor. The first layer convolves 32 (8 × 8) filters with stride 4 and uses a rectified linear unit (ReLU) activation function. The second layer is a batch normalization layer. The third layer convolves 64 (4 × 4) filters with stride 2 with a rectified linear unit (ReLU) activation function. The fourth layer is a batch normalization layer. The fifth layer is a fully-connected with 256 rectifier units. The sixth layer is fully-connected with 128 rectifier units and takes as input the concatenation of the output from the fifth layer and the action from the actor. The output layer is a fully-connected linear layer with a single output unit.
A major challenge of learning in continuous action spaces is exploration. An advantage of off-policy algorithms, such as DDPG, is that we can treat the exploration problem independently from the learning algorithm. We constructed an exploration policy μ by adding noise sampled from a noise process J to our actor policy:
μ ( s t ) = μ ( s t ) + J
where μ ( s t ) is the noised policy, μ ( s t ) is the current policy, and J is the action noise.
In the first published article based on DDPG and raw pixel images, the authors used the stochastic Ornstein–Uhlenbeck process [42] to generate random values temporally correlated as action noise. However, in the literature, we can also find implementations using exploratory noise from a normal distribution. Although these exploration approaches are proven to work, recent studies claim that parameter noise frequently boosts performance [43]. Parameter noise adds adaptive noise to the parameters of the NN policy (actor). It injects randomness directly on the weight of the NN, altering the type of actions the agent makes depending on what the agent currently senses. Different layers of the NN have different sensitivities to perturbation, which is why we add parameter noise to the last fully connected layers. In this paper, we will analyze the performance of the models using all three types of noise.
Finally, batches of 64 random samples are used from a replay buffer of size 25,000 to train the networks. Online actor and critic networks are trained by computing the target value and respective loss. The loss is the minimum squared error of the predicted and target values. The optimizer used is Adam. The actor and critic target networks are updated using soft target updates instead of directly copying the weights. Table 2 shows the values of the remaining hyperparameters.

5. Results and Discussion

This section presents the results obtained for both DQN and DDPG implementations, learning from raw pixel images to assign ATFM delay, with the dataset described in Section 4.1.

5.1. Performance Evaluation

The KPI is defined to evaluate the quality of the ATFM delay policy:
  • The sum of the rewards received by all the agents.
  • The sum of THE ATFM delay imposed by the agents.
  • The total number of delayed flights.
  • The sum of times the agents delayed a flight.
  • The mean OC of the sector along the episode.
These KPI’s are all relevant when evaluating the ATFM plan on a MARL system. One of the most widely used indicators to evaluate the performance of the agents is the sum of rewards earned at the end of each episode. The total delay imposed by the flights is also key because it is one of the indicators to minimize. The total number of delayed flights and the number of times the agents applied a delay (number of actions) can be considered KPIs showing how those delays are distributed among aircraft and the number of micro-adjustments agents make. The OC is key because it dictates situations with severe demand–capacity imbalances.
To compare the performance between the different implementations, Figure 3 shows the trend of the different KPIs using a moving window of fifty episodes. Those values have been obtained in all the cases, periodically testing the policy without exploratory noise.
The results demonstrate the potential of using RL algorithms based on images to solve DCB problems. As expected, the total reward per episode increases with the number of episodes, meaning that the agents are able to improve their policy by gathering experience from the environment. For the last 250 episodes, where we can assume convergence of the reward, DQN reported a reward of around –3000 while DDPG is around 1500. Note that the reward will always be smaller than 0 because the scenarios the agents will see always have DCB issues; thus, ATFM delay is mandatory. From the point of view of maximizing the cumulative reward, DDPG exhibits better performance than DQN.
The total ATFM delay shows a downward trend, denoting that the agents can infer which flights are more efficient to delay. DQN is the algorithm with the largest delay in the last episodes. DDPG with exploratory noise from a normal distribution reported the lowest delay while DDPG with Ornstein–Uhlenbeck and parameter noise reported an intermediate amount of delay. The main reason behind this difference in performance could come from the native characteristics of the algorithms. DQN uses discrete actions, which constrains the possible delay values, while DDPG uses continuous actions providing much more flexibility.
The number of delayed flights also decreases with similar behaviors between all the configurations, with an average value of around 20 delayed flights in the last 250 episodes. Although DQN and DDPG with Ornstein–Uhlenbeck seem to report slightly better performances, the improvements are minor.
Results related to the number of actions applied by the agents show that DQN is the algorithm with fewer micro-adjustments. DDPG with Ornstein–Uhlenbeck and parameter noise reported an intermediate similar number of actions. DDPG with noise from a normal distribution reported the highest value. This KPI is not directly linked to the goal of solving demand–capacity imbalances, but it is a good indicator of how many micro-adjustments are required to smooth the expected demand.
Related to the mean congestion of the sector, after 600 episodes, the sector’s mean OC seems to stabilize. DDPG with Ornstein–Uhlenbeck and parameter noise reports on average an 80% usage of the airspace sector capacity, while DDPG with normal noise and DQN exhibit around 90% usage of the capacity. As a reference, in the European ATM network, the desired occupancy value is around 80% of sector capacity, providing space to absorb unexpected events and ensuring that ATCOs are not overloaded [3].
Looking at the results of the different KPIs, it is not strongly clear which approach reports the best overall performance. While DDPG with normal noise excels at reducing the overall delay, DQN or DDPG with Ornstein–Uhlenbeck achieve a greater reduction in the number of affected flights, and DDPG with Ornstein–Uhlenbeck or DDPG with parameter noise further optimize the use of sector capacity. To better analyze the behavior of the algorithms from the DCB point of view, i.e., focusing on capacity usage to smooth the expected demand, Figure 4 shows the mean OC per episode for the DDPG implementations. The image shows the collected values per episode and their trend.
For the last 250 episodes where the mean OC converges, results from the DDPG with Ornstein–Uhlenbeck show the worst performance where many episodes reported a mean OC larger than the sustained capacity. DDPG with noise from a normal distribution reports slightly better results with fewer episodes with a demand greater than the sustained capacity on average. DDPG with parameter noise reports the best result with the smallest number of episodes with a mean demand larger than the sustained capacity.
Note that even though the algorithms do not keep the mean OC under the sustained capacity for all the episodes, for the last episodes where we can assume convergence in the performance, the mean demand does not exceed the peak capacity. Furthermore, focusing on the parameter noise implementation, it can be seen that the frequency and the number of consecutive episodes where the demand exceeds the sustained capacity are much smaller than in the other implementations.
Finally, a direct comparison between the results presented in this article and the actual ATFM delay is not feasible since the latter is the result of considering a broader environment. For example, let us imagine that a flight crosses two different regulated sectors. Even though the CASA algorithm could impose two different delays, the hypothetical flight would be affected only by the largest one. To directly compare the ATFM delay between the two approaches, a RL model for the two airspace sectors would be needed. Extending the proposed system to a broader region that considers the interaction between neighboring sectors is a relevant point to be studied in future works.

5.2. Case Study

This section presents the outcome of the framework for two specific regulations subtracted from the training dataset. The selected regulations are YBOLN07 from 7 September 2019, and YBOLN18A 18 August 2019. For each of the previous regulations, the RL system based on DDPG and parameter noise is used to collect which flights should be delayed and the amount of delay. Then, using this information, the original expected pre-tactical traffic is visualized using the following color schema:
  • Red: System-suggested flights for regulation.
  • Green: Non-regulated flights outside the sector in the corresponding timestamp.
  • Blue: Non-regulated flights inside the sector in the corresponding timestamp.
Figure 5 shows the results for regulation YBOLN07, which started at 8:00 A.M. and finished at 10:30 A.M. As a high-level indicator, the 141 flights crossing the sector linked to regulation YBOLN07 had a total delay of 556 min (delay from YBOLN07 or any other active regulation); thus, an average delay of 3.94 min per flight. On the other hand, our RL system suggests regulating 41 (from the 141 flights crossing the traffic volume) with an average delay per flight equal to 3.71 min per flight and a maximum individual delay equal to 21 min. Note that the comparison of minutes of delay per flight considers all the regulated traffic crossing the sector independently of the regulation.
Looking at the images, the selected sector (EDYYBOLN) has two traffic flows, one from the top-left to bottom-right and another from the bottom-left to top-right. Both traffic flows are similarly regulated, indicating that the delay is spread between fights, and the system does not have a preference. However, the RL policy sometimes decides to delay flights that do not completely cross the sector, which seems to be not ideal (see Figure 5, timestamp 8:57, red fight at the bottom-right).
Note that only five images at different timestamps per regulation are shown because of space constraints. Furthermore, there are no regulated flights in the first timestamp due to the problem constraint that must guarantee each episode starts without demand–capacity imbalance (see Section 3.5).
Figure 6 shows the results for regulation YBOLN18A, which started at 2:00 P.M. and finished at 4:45 P.M. In this case, the flights crossing the sector when the regulation was active received an ATFM delay of 3.39 min per flight, while the RL framework regulated 48 (from the 159 flights crossing the traffic volume) with an average delay per flight equal to 3.35 min per flight and a maximum individual delay equal to 18 min Notice that, despite the images being more crowded than in the previous case study, the average delay per flight is slightly smaller; 3.94 versus 3.39 for the actual ATFM delay and 3.71 versus 3.35 using the RL system. This is also the case for the peak delay imposed on individual flights.
The results obtained in these two case studies show the potential of the proposed new framework. The RL system is able to solve already detected DCB problems using images with a behavior that could be considered valid. However, a deeper analysis is required to obtain further conclusions.

6. Conclusions

This article proposes an image-based MARL solution to optimize ATFM delay in the European network. The goal is to maximize the usage of the airspace sector’s capacity while minimizing the ground delay. The proposed approach compares DQN and DDPG algorithms with the experience replay buffer, target networks, and different strategies for exploration. Although the obtained results did not lead to a clear conclusion about which algorithm configuration best fits the problem, DDPG arises as a promising candidate. It exhibits a lower overall ATFM delay and a mean OC closer to optimal values, especially if parameter noise is used for exploration during training.
The results obtained as a first step towards devising MARL methods for deciding on ATFM delay policies using pixel images are promising. The proposed system can successfully solve complex real-world DCB problems. Moreover, the work presented in this article could contribute to improving the usage of the airspace sector’s capacity and reducing current delays.
Another relevant aspect to highlight from this research is that the approach based on images for DCB problems provides a scalable architecture that allows the representation of an arbitrary number of agents without changing the state variables architecture. This characteristic is especially relevant when working on the entire European ATM network, where thousands of flights are operated daily.
As part of future work, it becomes mandatory to study how to extend the system to larger regions of multiple sectors, thus considering the effect that the decisions made in one sector generate in the neighborhood. Additional work towards a more realistic scenario would also include using the real polygonal shape of the sectors, considering more KPIs, such as cost, or analyzing the possibility of modeling heterogeneous agents. Finally, the combination of this work with the one presented in [7] will allow us to create an end-to-end system to precisely identify the airspace sectors with demand–capacity imbalances and propose a possible solution to reduce such demand during the pre-tactical phase.

Author Contributions

Conceptualization, S.M.-P. and E.S.; methodology, S.M.-P. and E.S.; software, S.M.-P.; validation, S.M.-P. and E.S.; formal analysis, S.M.-P.; investigation, S.M.-P. and E.S.; resources, E.S. and E.P.; data curation, S.M.-P.; writing—original draft preparation, S.M.-P.; writing—review and editing, S.M.-P. and E.S.; visualization, S.M.-P. and E.S.; supervision, E.S.; project administration, E.S.; funding acquisition, E.S. and E.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by EUROCONTROL under Ph.D. Research contract no. 18-220569-C2 and by the Ministry of Economy, Industry, and Competitiveness of Spain under grant number PID2020-116377RB-C21.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AIArtificial Intelligence
AIRACAeronautical Information Regulation and Control
ANSPAir Navigation Service Provider
ATCAir Traffic Control
ATCOAir Traffic Controller
ATFMAir Traffic Flow Management
ATMAir Traffic Management
CASAComputer Assisted Slot Allocation
CFMUCentral Flow Management Unit
CNNConvolutional Neural Network
CTOTCalculated Take-Off Time
DCBDemand–Capacity Balancing
DDPGDeep Deterministic Policy Gradient
DPGDeterministic Policy Gradient
DQNDeep Q-learning
EATMNEuropean Air Traffic Management Network
ECACEuropean Civil Aviation Conference
FMPFlow Manager Position
KPIKey Performance Indicator
MAA3CMulti-Agent Asynchronous Advantage Actor–Critic
MARLMulti-Agent Reinforcement Learning
MDPMarkov Decision Process
MLMachine Learning
MUACMaastricht Upper Area Control Centre
NNNeural Network
OCOccupancy Count
ReLURectified Linear Unit
RLReinforcement Learning
SOBTScheduled Off-Block Time

Nomenclature

α Learning rate
γ Discount factor
N Number of agents
μ DDPG ATFM delay policy
π DQN ATFM delay policy
aAction variable
CCapacity variable
DDelay variable
GCumulative reward
iAgent indicator
PTransition function
RReward variable
rReward variable
sState variable
T P Counting time period
VOccupancy Count variable

References

  1. Zhao, J.; Sun, J.; Cai, Z.; Wang, L.; Wang, Y. End-to-End deep reinforcement learning for image-based UAV autonomous control. Appl. Sci. 2021, 11, 8419. [Google Scholar] [CrossRef]
  2. Zhao, P.; Liu, Y. Physics informed deep reinforcement learning for aircraft conflict resolution. IEEE Trans. Intell. Transp. Syst. 2021, 23, 8288–8301. [Google Scholar] [CrossRef]
  3. ATFCM Users Manual 26.0. Available online: Available online: https://www.eurocontrol.int/publication/atfcm-users-manual (accessed on 2 August 2022).
  4. European Airline Delay Cost Reference Values—Updated and Extended Values—Version 4.1. Available online: https://www.eurocontrol.int/sites/default/files/publication/files/european-airline-delay-cost-reference-values-final-report-4-1.pdf (accessed on 2 August 2022).
  5. ATFM Modeling Capability. Available online: Available online: https://www.eurocontrol.int/sites/default/files/library/044_ATFM_Modelling_Capability.pdf (accessed on 2 August 2022).
  6. Performance Review Report (PRR) 2018. Available online: Available online: https://www.eurocontrol.int/publication/performance-review-report-prr-2018 (accessed on 2 August 2022).
  7. Mas-Pujol, S.; Salamí, E.; Pastor, E. RNN-CNN Hybrid Model to Predict C-ATC CAPACITY Regulations for En-Route Traffic. Aerospace 2022, 9, 93. [Google Scholar] [CrossRef]
  8. Mas-Pujol, S.; Salamí, E.; Pastor, E. Predict ATFCM weather regulations using a time-distributed Recurrent Neural Network. In Proceedings of the IEEE/AIAA 40th Digital Avionics Systems Conference, San Antonio, TX, USA, 3–7 October 2021; pp. 1–8. [Google Scholar]
  9. Dalmau, R.; Genestier, B.; Anoraud, C.; Choroba, P.; Smith, D. A machine learning approach to predict the evolution of air traffic flow management delay. In Proceedings of the 14th USA/Europe Air Traffic Management Research and Development Seminar (ATM2021), New Orleans, LA, USA, 20–24 September 2021. [Google Scholar]
  10. Jardines, A.; Soler, M.; García-Heras, J. Estimating entry counts and ATFM regulations during adverse weather conditions using machine learning. J. Air Transp. Manag. 2021, 95, 102109. [Google Scholar] [CrossRef]
  11. Ivanov, N.; Netjasov, F.; Jovanović, R.; Starita, S.; Strauss, A. Air Traffic Flow Management slot allocation to minimize propagated delay and improve airport slot adherence. Transp. Res. Part A Policy Pract. 2017, 95, 183–197. [Google Scholar] [CrossRef]
  12. Bolić, T.; Castelli, L.; Corolli, L.; Rigonat, D. Reducing ATFM delays through strategic flight planning. Transp. Res. Part E Logist. Transp. Rev. 2017, 98, 42–59. [Google Scholar] [CrossRef] [Green Version]
  13. Ruiz, S.; Kadour, H.; Choroba, P. An innovative safety-neutral slot overloading technique to improve airspace capacity utilisation. In Proceedings of the SESAR Innovation Days 2019, Athens, Greece, 2 December 2019. [Google Scholar]
  14. Dalmau, R. Predicting the likelihood of airspace user rerouting to mitigate air traffic flow management delay. Transp. Res. Part C Emerg. Technol. 2022, 144, 103869. [Google Scholar] [CrossRef]
  15. Delgado, L.; Prats, X. En route speed reduction concept for absorbing air traffic flow management delays. J. Aircr. 2012, 49, 214–224. [Google Scholar] [CrossRef]
  16. Fernández, E.C.; Cordero, J.M.; Vouros, G.; Pelekis, N.; Kravaris, T.; Georgiou, H.; Fuchs, G.; Andrienko, N.; Andrienko, G.; Casado, E.; et al. DART: A machine-learning approach to trajectory prediction and demand-capacity balancing. In Proceedings of the 11th SESAR Innovation Days, Online, 3–7 October 2021; pp. 1–10. [Google Scholar]
  17. Spatharis, C.; Bastas, A.; Kravaris, T.; Blekas, K.; Vouros, G.; Cordero, J.M. Hierarchical multiagent reinforcement learning schemes for air traffic management. Neural Comput. Appl. 2021, 1–13. [Google Scholar] [CrossRef]
  18. Huang, C.; Xu, Y. Integrated Frameworks of Unsupervised, Supervised and Reinforcement Learning for Solving Air Traffic Flow Management Problem. In Proceedings of the IEEE/AIAA 40th Digital Avionics Systems Conference, Online, 3–7 October 2021; pp. 1–10. [Google Scholar]
  19. Kravaris, T.; Vouros, G. Deep Multiagent Reinforcement Learning Methods Addressing the Scalability Challenge. In Multi-Agent Technologies and Machine Learning; IntechOpen Publishing House: London, UK, 2022; pp. 32–58. [Google Scholar]
  20. Tumer, K.; Agogino, A. Distributed agent-based air traffic flow managemen. In Proceedings of the 6th International Joint Conference on Autonomous Agents and Multiagent Systems, New York, NY, USA, 14 May 2007. [Google Scholar]
  21. Crespo, A.; Marcio, F.; Weigang, L.; Barros, A.G. Reinforcement learning agents to tactical air traffic flow management. Int. J. Aviat. Manag. 2012, 1, 145–161. [Google Scholar] [CrossRef]
  22. Montague, P.R. Reinforcement learning: An introduction, by Sutton, R.S. and Barto, A.G.; MIT Press: Cambridge, MA, USA, 1999; Volume 3. [Google Scholar]
  23. Watkins, C.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  24. Tan, M. Multi-agent reinforcement learning: Independent vs. cooperative agents. In Proceedings of the Tenth International Conference on Machine Learning, Amherst, MA, USA, 27–29 July 1993. [Google Scholar]
  25. Lowe, R.; Wu, Y.I.; Tamar, A.; Harb, J.; Pieter, A.; Pieter Abbeel, O.; Mordatch, I. Multi-agent actor-critic for mixed cooperative-competitive environments. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  26. Foerster, J.; Farquhar, G.; Afouras, T.; Nardelli, N.; Whiteson, S. Counterfactual multi-agent policy gradients. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  27. Sunehag, P.; Lever, G.; Gruslys, A.; Czarnecki, W.M.; Zambaldi, V.; Jaderberg, M.; Lanctot, M.; Sonnerat, N.; Leibo, J.Z.; Tuyls, K.; et al. Value-decomposition networks for cooperative multi-agent learning. arXiv 2017, arXiv:1706.05296. [Google Scholar]
  28. Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. Deep reinforcement learning: A brief survey. IEEE Signal Process. Mag. 2017, 34, 26–38. [Google Scholar] [CrossRef] [Green Version]
  29. Silver, D.; Lever, G.; Heess, N.; Degris, T.; Wierstra, D.; Riedmiller, M. Deterministic policy gradient algorithms. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014. [Google Scholar]
  30. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  31. Hafner, R.; Riedmiller, M. Reinforcement learning in feedback control. Mach. Learn. 2011, 84, 137–169. [Google Scholar] [CrossRef] [Green Version]
  32. Basora, L.; Morio, J.; Mailhot, C. A trajectory clustering framework to analyse air traffic flows. In Proceedings of the 7th SESAR Innovation Days, Belgrade, Serbia, 28–30 November 2017; pp. 1–8. [Google Scholar]
  33. Corrado, S.J.; Puranik, T.G.; Pinon, O.J.; Mavris, D.N. Trajectory clustering within the terminal airspace utilizing a weighted distance function. Proceedings 2020, 59, 7. [Google Scholar] [CrossRef]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef] [Green Version]
  35. Egorov, M. Multi-agent deep reinforcement learning. Convolutional Neural Netw. Vis. Recognit. 2016, 10, 1–8. [Google Scholar]
  36. Agogino, A.; Tumer, K. Learning indirect actions in complex domains: Action suggestions for air traffic control. Adv. Complex Syst. 2009, 12, 493–512. [Google Scholar] [CrossRef]
  37. Spatharis, C.; Kravaris, T.; Vouros, G.A.; Blekas, K.; Chalkiadakis, G.; Garcia, J.M.C.; Fernandez, E.C. Multiagent reinforcement learning methods to resolve demand capacity balance problems. In Proceedings of the 10th Hellenic Conference on Artificial Intelligence, New York, NY, USA, 9–12 July 2018. [Google Scholar]
  38. Cruciol, L.; Arruda, A., Jr.; Weigang, L.; Crespo, A. Reward functions for learning to control in air traffic flow management. Transp. Res. Part C Emerg. Technol. 2013, 35, 141–155. [Google Scholar] [CrossRef]
  39. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. Adv. Neural Inf. Process. Syst. 2014, 2224–2232. [Google Scholar] [CrossRef]
  41. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  42. Uhlenbeck, G.E.; Ornstein, L.S. On the theory of the Brownian motion. Phys. Rev. 1930, 36, 823. [Google Scholar] [CrossRef]
  43. Plappert, M.; Houthooft, R.; Dhariwal, P.; Sidor, S.; Chen, R.Y.; Chen, X.; Asfour, T.; Abbeel, P.; Andrychowicz, M. Parameter space noise for exploration. arXiv 2017, arXiv:1706.01905. [Google Scholar]
Figure 1. Interaction between the different elements in a RL system (adapted from [23]).
Figure 1. Interaction between the different elements in a RL system (adapted from [23]).
Aerospace 09 00599 g001
Figure 2. Three channels, image-like, represent the input states of the RL system. The inside and outside channels encode information about the flights. The self-channel encodes information about the agent that is taking action. The inner square represents the shape of the sector.
Figure 2. Three channels, image-like, represent the input states of the RL system. The inside and outside channels encode information about the flights. The self-channel encodes information about the agent that is taking action. The inner square represents the shape of the sector.
Aerospace 09 00599 g002
Figure 3. Trends KPIs used to evaluate the performances of the RL systems.
Figure 3. Trends KPIs used to evaluate the performances of the RL systems.
Aerospace 09 00599 g003
Figure 4. Mean occupancy count per episode for the DDPG implementations. (a) Ornstein–Uhlenbeck noise, (b) normal distribution noise, and (c) parameter noise. The purple line shows the trend, the red line represents the sustained capacity, and the green line represents 80% of the sector’s capacity.
Figure 4. Mean occupancy count per episode for the DDPG implementations. (a) Ornstein–Uhlenbeck noise, (b) normal distribution noise, and (c) parameter noise. The purple line shows the trend, the red line represents the sustained capacity, and the green line represents 80% of the sector’s capacity.
Aerospace 09 00599 g004
Figure 5. Representation of the RL system outcome for the regulation YBOLN07.
Figure 5. Representation of the RL system outcome for the regulation YBOLN07.
Aerospace 09 00599 g005
Figure 6. Representation of the RL system outcome for the regulation YBOLN18A.
Figure 6. Representation of the RL system outcome for the regulation YBOLN18A.
Aerospace 09 00599 g006
Table 1. Hyperparameters for the deep Q-learning algorithm.
Table 1. Hyperparameters for the deep Q-learning algorithm.
HyperparameterValueDescription
Episode1000Total number of training episodes
Max steps60Maximum number of steps per episode
Number of actions4Number of different actions
Discount factor0.99Discount factor of future rewards
Learning rate0.00025Learning rate used by the optimizer
Initial ϵ 1Initial value for exploration
Final ϵ 0.1Minimum value for exploration
Target update4Step frequency to update the target network
Table 2. Hyperparameters for the DDPG algorithm.
Table 2. Hyperparameters for the DDPG algorithm.
HyperparameterValueDescription
Episode1000Total number of training episodes
Max steps60Maximum number of steps per episode
Discount factor0.99Discount factor of future rewards
Learning rate actor0.001Learning rate used by the optimizer
Learning rate critic0.002Learning rate used by the optimizer
Initial ϵ 1Initial value for exploration
Final ϵ 0.1Minimum value for exploration
Target update4Step frequency to update the target network
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mas-Pujol, S.; Salamí, E.; Pastor, E. Image-Based Multi-Agent Reinforcement Learning for Demand–Capacity Balancing. Aerospace 2022, 9, 599. https://doi.org/10.3390/aerospace9100599

AMA Style

Mas-Pujol S, Salamí E, Pastor E. Image-Based Multi-Agent Reinforcement Learning for Demand–Capacity Balancing. Aerospace. 2022; 9(10):599. https://doi.org/10.3390/aerospace9100599

Chicago/Turabian Style

Mas-Pujol, Sergi, Esther Salamí, and Enric Pastor. 2022. "Image-Based Multi-Agent Reinforcement Learning for Demand–Capacity Balancing" Aerospace 9, no. 10: 599. https://doi.org/10.3390/aerospace9100599

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop