Next Article in Journal
Object Detection in Adverse Weather for Autonomous Driving through Data Merging and YOLOv8
Previous Article in Journal
Peptide-Functionalized Carbon Nanotube Chemiresistors: The Effect of Nanotube Density on Gas Sensing
Previous Article in Special Issue
A Novel Feeder Link Handover Strategy for Backhaul in LEO Satellite Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Hierarchical Admission Control for Low-Earth Orbit Satellites Based on Deep Reinforcement Learning

1
Communication and Network Laboratory, Dalian University, Dalian 116622, China
2
School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(20), 8470; https://doi.org/10.3390/s23208470
Submission received: 31 August 2023 / Revised: 1 October 2023 / Accepted: 3 October 2023 / Published: 14 October 2023
(This article belongs to the Special Issue Satellite Based IoT Networks for Emerging Applications)

Abstract

:
Low-Earth orbit (LEO) satellites have limited on-board resources, user terminals are unevenly distributed in the constantly changing coverage area, and the service requirements vary significantly. It is urgent to optimize resource allocation under the constraint of limited satellite spectrum resources and ensure the fairness of service admission control. Therefore, we propose an intelligent hierarchical admission control (IHAC) strategy based on deep reinforcement learning (DRL). This strategy combines the deep deterministic policy gradient (DDPG) and the deep Q network (DQN) intelligent algorithm to construct upper and lower hierarchical resource allocation and admission control frameworks. The upper controller considers the state features of each ground zone and satellite resources from a global perspective, and determines the beam resource allocation ratio of each ground zone. The lower controller formulates the admission control policy based on the decision of the upper controller and the detailed information of the users’ services. At the same time, a designed reward and punishment mechanism is used to optimize the decisions of the upper and lower controllers. The fairness of users’ services admissions in each ground zone is achieved as far as possible while ensuring the reasonable allocation of beam resources among zones. Finally, online decision-making and offline learning were combined, so that the controller could make full use of a large number of historical data to learn and generate intelligent strategies with stronger adaptive ability while interacting with the network environment in real time. A large number of simulation results show that IHAC has better performance in terms of a successful service admission rate, service drop rate, and fair resource allocation. Among them, the number of accepted services increased by 20.36% on average, the packet loss rate decreased by 17.56% on average, and the resource fairness increased by 17.16% on average.

1. Introduction

With the continuous advancement of communication technologies and the rapid growth of global communication demands, the existing terrestrial network can provide high-speed and diversified services for densely populated areas, but it still has obvious shortcomings in dealing with long distances, a wide coverage, large capacity, and harsh terrain communication environment. At the same time, as an important supplement to the ground network, a satellite network has significant advantages in expanding the communication coverage, improving the communication capacity and reliability in hot spots, and realizing the space–air–ground–sea integrated network [1,2]. Among them, low-earth orbit (LEO) satellite communication systems have become a project pursued by many enterprises in the Internet, communication, aerospace, and other fields due to its advantages, including a low delay, low cost, high spectrum utilization, and low terminal power requirements [3]. However, LEO satellites are characterized by their low cost, small size, and lightweight design, but they result in severely limited on-board power resources. LEO satellites orbit at low altitudes and move at high speeds, leading to constantly changing coverage areas and complex and dynamic electromagnetic environments. Furthermore, user terminals are unevenly distributed across different countries and regions within the coverage area, and there are varying demands for traditional voice and data services as well as emerging streaming media services. In summary, LEO satellites must address the dynamic changes in user terminal distribution and service demands. Hence, addressing the challenge of allocating resources effectively within the constraints of limited resources such as the satellite spectrum, ensuring equitable admission control for diverse services, and achieving efficient alignment between on-board resources and service demands has emerged as a pressing issue in the realm of satellite communication.
Multi-beam satellite communication technology is an effective approach to solve the above challenges. This technology achieves multiple frequency reuses and polarization reuse through spatial isolation, which can increase satellite communication capacity exponentially. At the same time, the dynamically adjusted beam direction can help the satellite system allocate resources more effectively and further meet the needs of the variability in coverage area and the diversity of communication requirements [4]. However, the flexibility of multi-beam satellite communication systems also increases the complexity of resource allocation. There are usually different communication requirements in the coverage area of each beam, which requires the satellite communication system to be able to sense the differences between beams and dynamically adjust the resource allocation scheme to ensure that the services within each beam are properly served [5]. For these reasons, the academic community has proposed some solutions from the perspectives of beamforming schemes and on-board resource joint allocation.
Lin et al. [6,7] proposed a beamforming scheme for a multi-beam satellite communication system from the perspective of security and energy saving, thereby improving the secrecy energy efficiency (SEE) of communication networks. Lin et al. [8] adopted the alternating optimization scheme and the Taylor expansion penalty function to optimize the beamforming weight vector and phase shift, aiming to minimize the total transmission power of the satellite and the base station, while meeting the user rate requirements. Deng et al. [9] proposed an adaptive packet splitting scheme based on a discrete firefly algorithm for cross-layer and cross-dimension radio resources optimization and an irregular gradient algorithm to ensure communication efficiency and reliability. Zhang et al. [10] proposed an uplink cooperative user scheduling and power allocation method based on game theory in an uplink multi-beam satellite Internet of Things (S-IoT). Takahashi et al. [11] constructed a power resource allocation model to jointly control the transmit power and multi-beam directivity according to the traffic demand, so as to improve the efficiency of communication resource allocation. Jia et al. [12] considered the inter-beam interference, channel conditions, delay, capacity, bandwidth utilization variance, and other factors, and a joint resource allocation algorithm is proposed, which can flexibly allocate resources according to specific service requirements and channel conditions. Wang et al. [13] proposed a resource management scheme to maximize resource utilization and user service weight, and introduced an improved cuckoo optimization algorithm to ensure the quality of service (QoS) for high-priority users. In addition, Cai et al. [14] proposed a cognitive S-IoT system supporting code division multiple access (CDMA), which can jointly optimize the transmission power of the traditional satellite system and the IoT by maximizing the total rate of the IoT users under the premise of ensuring the performance requirements of the traditional satellite system.
The emergence of the above research results brings new ideas and methods for the resource allocation of multi-beam LEO satellite communication. However, when dealing with the diverse and time-varying communication requirements in the network, these methods are difficult for meeting the needs of rapid and dynamic resource allocation due to high algorithm complexity and long calculation times. Machine learning (ML) [15] has super strong learning and inference capabilities, and has been shown as a good application prospect in network resource allocation [16,17,18,19]. In particular, the introduction of deep reinforcement learning (DRL) provides a more intelligent, flexible, and efficient solution for the resource allocation of multi-beam LEO satellite communication.
Liao et al. [20] designed a cooperative multi-agent deep reinforcement learning (CMDRL) framework to solve the bandwidth allocation problem of multi-beam satellite communication systems. Chan et al. [21] proposed an efficient power and bandwidth allocation method, which uses two linear ML algorithms and takes channel conditions and traffic demand as inputs to achieve the optimal resource allocation under dynamic channel conditions. Hu et al. [22] considered inter-beam interference and resource utilization difference, and a game theory based on bandwidth allocation model for forward links was established. Furthermore, considering that the addition of satellite beams results in a larger action space for individual agents in DRL, thereby increasing time complexity, this paper proposed a multi-agent cooperative deep reinforcement learning approach to achieve optimal bandwidth allocation. He et al. [23] proposed a multi-objective deep-reinforcement-learning-based time-frequency (MODRL-TF) two-dimensional resource allocation algorithm to achieve the joint optimization goal of maximizing the number of users and system throughput for the joint allocation problem of multi-dimensional resources such as QoS, time, and frequency in multi-beam satellite communication. Huang et al. [24] proposed a learning-based hybrid-action deep Q-network (HADQN) algorithm to solve the sequential decision optimization problem in dynamic resource allocation in multi-beam satellite systems. By using a parameterized hybrid action space, HADQN can schedule the beam pattern and allocate the transmitter power more flexibly, which greatly reduces the on-orbit energy consumption without affecting the QoS.
By introducing DRL, the above research provided a forward-looking solution for solving resource allocation in multi-beam LEO satellite networks. However, current research still has some shortcomings, with the main deficiency being the lack of consideration for fair user service admission in the optimal allocation of resources. In the multi-beam LEO satellite communication system, the coverage area of the satellite is constantly changing, and the needs of users in the area are diverse.
If fairness is not considered in resource allocation, this will result in some users being overlooked or frequently denied access, while others consistently occupy more resources. Over time, this leads to a significant decline in user experience, causing dissatisfaction with the services provided by the satellite. However, in the existing research, traditional algorithms exhibit limited flexibility and adaptability when confronted with the complexities of the satellite communication environment. Moreover, accurately extracting the key information required for fair admission control strategies from a vast array of service demand characteristics often proves challenging. Additionally, a comprehensive and effective hierarchical admission control strategy capable of achieving the cooperative optimization of resource allocation rationality and admission control fairness has yet to be established. Hence, we propose an intelligent hierarchical admission control (IHAC) strategy based on DRL. This approach incorporates the user service model, the dynamic priority model, and carefully designed controller feature inputs, along with a reward and punishment mechanism. Through the utilization of DRL, IHAC aims to efficiently allocate multi-beam resources and exercise control over user service access within the coverage area, simultaneously addressing both objectives.
The contributions of this paper are as follows:
  • In this paper, the DLR method is employed to effectively manage the multi-beam satellite communication system. By combining deep deterministic policy gradient (DDPG) and deep Q-network (DQN) algorithms, a hierarchical admission control strategy is constructed, encompassing both upper- and lower-level controllers. The upper-level controller employs DDPG to comprehensively consider the service demands of various ground zones and formulate resource scheduling strategies from a global perspective. This ensures optimized resource utilization while ensuring the coordinated allocation of beam resources among different ground zones. The lower-level controller employs DQN to receive decisions from the upper-level controller and refines the global strategy into localized admission control strategies for each ground zone’s services.
  • This paper develops user service models and dynamic priority models for services. The user service model enables controllers to have a deeper understanding of different types of service characteristics, leading to more precise decision-making. By introducing the adaptive dynamic priority model, controllers can achieve efficient resource utilization while considering the fairness of competitive services, leading to a more balanced resource allocation. Furthermore, careful design is applied to the feature inputs, reward functions, and penalty terms of both upper- and lower-level controllers, facilitating a more comprehensive and detailed understanding of service demands. This guides controller decisions towards optimizing both efficient resource utilization and overall fairness of the communication system.
  • In simulation experiments, the proposed strategy is compared with DAQC [25] and WOA [26]. The results demonstrate that the proposed strategy performs well in terms of the channel access service quantity, successful service admission rate, service drop rate, and resource allocation fairness. This highlights the effectiveness of the proposed strategy.
The remainder of this paper is organized as follows. Section 1 describes the related works on resource allocation and admission control for multi-beam satellites and the motivation of this paper. Section 2 describes the system model of the resource allocation and admission control strategy in this paper. Section 3 outlines the hierarchical admission control framework based on DRL. Section 4 provides a detailed explanation of the hierarchical admission control strategy. Section 5 presents the evaluation results of the experiments. Section 6 summarizes the entire paper and elaborates on the contributions of this work.

2. System Model

2.1. Communication Model of the Satellite Coverage Area System

In LEO satellite communication systems, as the satellite moves in the orbit, its coverage will change, and the corresponding traffic load and service priorities will also change. In order to meet the changing requirements of ground tasks, it is necessary to allocate satellite resources reasonably. The resource allocation problem is mainly affected by time window constraints, satellite power consumption constraints, ground user priorities, etc.
The system model for multibeam LEO satellite communication is illustrated in Figure 1. We simplify the satellite coverage area into rectangles and further divide it into M fixed ground zones, and the communication tasks are assigned to different ground zones according to the location of their initiating users. In the context of multibeam satellite communication, LEO satellites allocate their transmission capacity into multiple beams, each directed towards a specific ground zone. LEO satellites can dynamically adjust the beam resources of ground zones in real-time based on the distribution of services on the ground, and further decide whether to admit user services within these zones. This communication architecture between user terminals and satellites forms a complex system.

2.2. User Service Model

The core focus of this paper is to address the complex issue of user service requests within multibeam satellite communication systems. The initial step involves constructing a user service model to depict various types of user services. This user service model comprehensively accounts for various parameters and features of the requested services to ensure a comprehensive and accurate representation of user service requirements. In addition to the user model, we also establish a dynamic user service priority model. Given the constantly changing communication environment, service priorities are adjusted based on dynamically changing conditions, such as the remaining service value and service urgency. By introducing an adaptive dynamic priority mechanism, the user service admission process is further optimized to ensure efficient resource utilization, while achieving fair resource allocation among competitive services. Among them, the relevant variable descriptions of user services are presented in Table 1.
With the movement of satellites and the progression of time, the status of user services is constantly changing. We design a sextuple to represent the data model of user services within a beam. Let there be M ground zones within the satellite coverage rectangular area, each containing N user services. To account for the service status of the satellite beam, the user services U m , n within the beam coverage area are as follows.
U m , n = ( B m , n , P m , n , C m , n , V m , n , T s m , n , S m , n ) .
where m { 1 , 2 , , M } , n { 1 , 2 , , N } ; B m , n and P m , n satisfy the condition m = 1 M m = 1 N B m , n B t o t a l , m = 1 M n = 1 N P m , n P t o t a l ; the initial priority V m , n is from the set { 1 , 2 , 3 , 4 , 5 } , corresponding to background flow service, interactive service, streaming media service, audio service, and handover service; S m , n satisfies the condition S T T s m , n E T , S m , n E T S T .
Building upon the initial priority of user services, this paper also takes into account the aspects of the remaining service value and the urgency of service provisioning. The dynamic priority D P m , n is defined as:
D P m , n = V m , n × t t E T t + 1
Let t s t a r t be the time that the service be served. If the service has been served, then t = t s t a r t ; if the service has not been served, then t = T s m , n , and S m , n E T t . According to Equation (2), the dynamic priority of the user service is directly proportional to the initial priority of the service and the time for the service to be served, and inversely proportional to the remaining time of this resource allocation time window. The dynamic priority introduced in this paper enables the strategy to comprehensively understand and evaluate service priorities, leading to more intelligent service management and admission decisions.

3. Framework of Intelligent Hierarchical Admission Control Based on DRL

3.1. Overview

DRL, as a powerful artificial intelligence technique, offers the core advantage of autonomous learning and optimization in complex and uncertain environments. It achieves this by interacting with the environment, accumulating experiences, and progressively improving decision quality [27,28,29,30].
The integration of the DDPG algorithm and DQN algorithm, which are prominent components of DRL, provides a novel solution for hierarchical admission control in satellite communication systems. DDPG extends the concepts of DQN and the Actor-Critic framework, effectively handling the complexity of state spaces, making it particularly suitable for the continuous action control problem of allocating beam resources to ground zones. On the other hand, DQN can accurately perceive and analyze the state features and communication demands of user services, addressing the discrete admission control problem. Combining these DRL algorithms with LEO satellite communication hierarchical admission control leads to intelligent decisions regarding user admission into the communication network and optimal resource allocation under limited resource conditions.
The combination of the above DRL algorithms, LEO satellite communication resource allocation, and admission control forms a two-level intelligent admission control framework, as illustrated in Figure 2. In this framework, the upper-level controller bears the responsibility of intelligent decision-making on beam resource allocation for ground zones based on their status features and communication demands. Meanwhile, the lower-level controller focuses on the admission control of user services within each zone. Guided by upper-level decisions and considering the actual situations of user services in each zone, the lower-level controller manages communication resources more delicately, ensuring a certain level of fairness in the allocation of user service resources. Among them, the parameters in the DRL modeling process are defined in Table 2.
Within the context of this hierarchical admission control strategy, it is necessary to holistically consider the relationships among all ground zones, while also giving due consideration to the priority and fairness management among user services. The implementation of this strategy introduces a new level of intelligence to the entire communication system, allowing the system to better adapt to diverse communication needs and provide users with a higher-quality communication experience.

3.2. DRL Formulation

3.2.1. State

As illustrated in Figure 2, the upper-level and lower-level controllers are respectively equipped with DDPG and DQN neural networks, with DDPG further divided into Actor and Critic neural networks. The neural networks are required to make decisions based on the state features of satellite communication. To enhance the neural networks’ perception capabilities concerning ground zones, user service states, and communication demands, ensuring that the admission control strategy maintains fairness across various service resource allocations, we have specially designed the input features for the upper- and lower-level neural networks. These input features are crafted to provide richer information, enabling the neural networks to make more precise decisions across diverse scenarios, thereby effectively optimizing the performance and resource allocation of the satellite communication system.
In the upper-level admission control strategy, each ground zone initially conducts unified scheduling and management of its internal services. After analyzing and processing user service information, the ground zone’s state features are formed and submitted to the upper-level DDPG. The input features encompass both user service and ground zone information, such as the number of service requests in the ground zone, total dynamic priority, and aggregate requested bandwidth. This empowers the DDPG to gain comprehensive insights into interactions and communication needs among users within the ground zone, thus providing more precise guidance for higher-level decisions. Meanwhile, the input features for the lower-level DQN include the upper-level DDPG’s resource allocation strategy for the ground zone and more detailed user service information, such as service types. This design enables the lower-level DQN to more accurately assess each user’s communication performance, enabling more targeted admission control decisions.
To be specific, the upper-level DDPG’s input features are denoted as s t u p = ( s t , 1 u p , s t , 2 u p , , s t , m u p , , s t , M u p , s t , B u p , s t , P u p ) , where s t , m u p signifies the state features of the mth ground zone, and s t , B u p and s t , P u p represent the satellite’s total bandwidth and total power, respectively. The lower-level DQN’s input features are represented as s t , m d o w n .
s t , m u p = ( N m , P r m , W t m , B m , P m , C m ) .
s t , m d o w n = ( B ^ m , n , P ^ m , n , C ^ m , n , V m , n , T s m , n , S m , n , D P m , n , a t , m u p ) .
The specific features are given by the following formulas.
P r m = n = 1 N m D P m , n .
W t m = n = 1 N m ( t T s m , n ) .
B m = n = 1 N m B m , n .
P m = n = 1 N m P m , n .
C m = n = 1 N m B m , n   log 2 ( 1 + S I N R m ) .
S I N R m represents the signal-to-interference-plus-noise Rrtio (SINR) of the mth beam:
S I N R m = A m n = 1 N m P m , n N 0 n = 1 N m B m , n + i M , i m A i n = 1 N i P i , n .
where A m is the channel attenuation coefficient of the mth ground zone; N 0 is the power spectral density of additive white Gaussian noise; i M , i m A i n = 1 N i P i , n is the sum of co-channel interference from other beams to the m th beam.
The ratios of the requested bandwidth, power, and capacity of the n th user service to the total requested resources in the m th ground zone B ^ m , n , P ^ m , n , and C ^ m , n are as follows:
B ^ m , n = B m , n / B m , P ^ m , n = P m , n / P m , C ^ m , n = C m , n / C m .

3.2.2. Action

In the proposed hierarchical admission control architecture, the upper- and lower-level controllers are responsible for the scheduling of satellite-to-ground beam resources and the admission control of user services within ground zones, respectively. The upper-level controller plays a crucial role in global decision-making. It aggregates network state features from various ground zones and feeds this information into the DDPG neural network. Subsequently, based on these network state features, DDPG formulates a set of resource scheduling policies a t u p = ( a t , 1 u p , a t , 2 u p , , a t , m u p , , a t , M u p ) for each ground zone, where a t , m u p [ 0 , 1 ) , m = 1 M a t , m u p = 1 , and a t , m u p represent the scheduling policy for the m th ground zone’s beam resources.
Specifically, when a t , m u p ( 0 , 1 ) , it indicates that the satellite decides to admit the m th ground zone and allocate satellite resources to the beams of the current ground zone according to a certain percentage. On the other hand, when a t , m u p = 0 , it signifies that the satellite will not admit that ground zone.
Once the satellite completes the resource scheduling for various ground zones, it enters the domain of the lower-level controller. The primary responsibility of the lower-level controller is the admission control of user services within each ground zone. At this stage, the lower-level DQN leverages the state feature s t , m d o w n of user services within its respective ground zone to formulate admission control policies a t , m d o w n { 0 , 1 } for each user service. When a t , m d o w n = 1 , it signifies that the corresponding user service is admitted, whereas a t , m d o w n = 0 indicates that the user service is not admitted.

3.2.3. Reward

In the proposed intelligent hierarchical admission control framework based on DRL, the reward function plays a crucial role. The design of this key element not only establishes a close correlation between the network state feature input and the admission control policy output but also guides the neural network towards rapidly approaching and converging to the predefined optimization objectives. During the resource scheduling process for various ground zones by the satellite, the design of the reward function for the upper-level DDPG needs to take into account the overall system performance, ensuring both communication system efficiency and fairness. This prevents excessive bias of communication resources towards specific ground zones, which could impact the overall communication effectiveness. Therefore, the reward function r t u p for the upper-level DDPG is formulated as follows:
r t u p = α 1 m = 1 M S P R m S P R a v e 2 M + α 2 m = 1 M D W T m D W T a v e 2 M .
where α 1 ,   α 2 are weight coefficients that weigh the importance of two terms in (12). The reward function r t u p comprehensively evaluates the fairness of the satellite’s admission control strategy a t u p for various ground zones by considering the variance of the total dynamic priority of successfully admitted services 1 M m = 1 M S P R m S P R a v e 2 and the variance of the total waiting time of services not admitted across all ground zones 1 M m = 1 M D W T m D W T a v e 2 .
The lower-level DQN is built upon the foundation of the upper-level admission control strategy and aims to make decisions for services within each ground zone. In this architecture, a reasonable lower-level admission control strategy’s primary concern is to ensure compliance with the resource allocation proportions set by the upper level. If the total resource consumption of admitted services exceeds these limits, the lower-level DQN reward function needs to impose a certain penalty to maintain system balance and stability. Similarly, the fairness of the intra-zone service admission control strategy must also be ensured. Therefore, we employ the total dynamic priority S P R m , the total waiting time of successfully admitted services S W T m , and the reward or penalty R P m for the resource utilization rate of the ground area admission control strategy to design the reward function r t u p for the lower-level DQN:
r t , m d o w n = β 1 S P R m + β 2 S W T m + β 3 R P m
To ensure data consistency, we have normalized these values. Through these two components, the lower-level DQN is guided to balance the dynamic priority and waiting time of intra-zone user services. Additionally, R P m is obtained as follows:
R P m = 500 , R U m = 0 10 log 1.1 ( R U m ) , 0 < R U m < 1 10 , R U m = 1 10 log 0.9 ( R U m ) , R U m > 1 .
where R U m represents the ratio of resources allocated to user services within the m th ground zone to the resources allocated by the upper-level control strategy.

4. Strategy of Intelligent Hierarchical Admission Control Based on DRL

4.1. Hierarchical Controller Structure

We delve into the exploration of DDPG and DQN, evolved from the traditional Q-learning algorithm, and their collaborative implementation in the hierarchical admission control framework. However, these methods also inherit issues such as instability, overestimation, and convergence challenges that have been associated with traditional Q-learning, especially when utilizing neural networks to estimate Q-values.
In conventional Q-learning, a common practice involves using a single online network to estimate Q-values for each state-action pair, and decisions are made based on the action with the highest Q-value. Nevertheless, this approach can lead to unstable Q-value estimations, particularly during the initial stages of training. This instability arises from the chain reaction triggered by the parameter updates in the neural network, resulting in fluctuations in Q-values. Furthermore, as traditional Q-learning employs a greedy strategy to select actions with the highest Q-values, it can lead to the overestimation of Q-values, subsequently affecting the accuracy of the policy.
To overcome these challenges, we introduce the concepts of a target network and experience replay mechanism in DRL. Specifically, the target network shares the same structure as the online network but employs a different parameter update strategy. The target network’s parameters are copied from the online network at regular intervals or through a soft update approach. This separation of neural networks ensures that the calculation of the target network Q-values is unaffected by the direct influence of the online network, reducing the immediate impact of the online network’s instability and mitigating the chain reaction caused by parameter updates.
Furthermore, in interactive environments, the acquired data are often highly correlated, with potential correlations between consecutive data points. This correlation can lead to the neural network being influenced by related data during the training process, thereby causing instability. The experience replay mechanism, by randomly sampling from an experience buffer, breaks down data correlation, enhancing data utilization efficiency. The deployment of neural networks is as follows.

4.1.1. DDPG in the Upper-Level Controller

Online Actor Network: π ( s t u p | θ ) , where θ represents the neural network parameters. Responsible for generating the scheduling strategy a t u p for various ground zones’ beam resources based on the ground zone and satellite resource state features s t u p .
Online Critic Network: Q ( s t u p , a t u p | ω ) , where ω represents the neural network parameters. Primarily evaluates the quality of actions a t u p made by the online actor network based on s t u p , guiding the optimization of the action network’s decision-making capability.
Target Actor Network and Target Critic Network: π ( s t + 1 u p | θ ) and Q ( s t + 1 u p , a t + 1 u p | ω ) , where θ and ω are the parameters of these two neural networks, and a t + 1 u p is the action computed by the target action network based on s t + 1 u p , but not executed during the interaction process.

4.1.2. DQN in the Lower-Level Controller

Online DQN: Q ( s t , m d o w n , a t , m d o w n | μ ) , where μ represents the neural network parameters. In the distributed scenario of ground zone user service admission control, each ground zone possesses an independent online DQN, sequentially making admission control decisions for the user service within its zone.

4.2. DRL Training

While interacting with the environment, the upper- and lower-level controllers accumulate a variety of state-action pairs’ experience samples ( s t u p , a t u p , s t + 1 u p , r t u p ) and ( s t , m d o w n , a t , m d o w n , s t + 1 , m d o w n , r t , m d o w n ) by trying different actions. These experience samples encompass the behavior and outcomes of the satellite communication system under various states. They are stored in an experience replay pool for neural network training. By obtaining N ( s i u p , a i u p , s i + 1 u p , r i u p ) and N ( s i , m d o w n , a i , m d o w n , s i + 1 , m d o w n , r i , m d o w n ) through random sampling, the correlation between samples is broken, thereby enhancing training stability and the system’s understanding and adaptability to different situations.

4.2.1. Training of the Upper-Level Controller DDPG

During the training phase of DDPG, the online actor network utilizes the experience samples N ( s i u p , a i u p , s i + 1 u p , r i u p ) from the replay pool to update its weights based on Equation (15):
θ J 1 N i = 1 N a Q ( s i u p , a i u p | ω ) θ π ( s i u p | θ ) .
The online actor network relies on the backpropagation of the policy gradient θ J for updates. Additionally, the online critic network can be trained by minimizing the loss function, as expressed below:
L o s s ( ω ) = 1 N i N ( y i u p Q ( s i u p , a i u p | ω ) ) 2 ,
where y i u p is referred to as the target Q-value, which can be expressed in a sample ( s i u p , a i u p , s i + 1 u p , r i u p ) as follows:
y i u p = r i u p + γ Q ( s i + 1 u p , a i + 1 u p | ω ) ,
The target Q-value is obtained by taking a weighted sum of the current reward r i u p and Q ( s i + 1 u p , a i + 1 u p | w ) , where γ is considered the discount factor, and a i + 1 u p is computed by the target action network, expressed as follows:
a i + 1 u p = π ( s i + 1 u p | θ ) .
Finally, we adopted a soft update approach for updating the parameters of both target networks, where in each update iteration, the parameters are updated by weighting them with the learning rate τ , expressed as follows:
ω = τ ω + ( 1 τ ) ω , θ = τ θ + ( 1 τ ) θ .

4.2.2. Training of the Lower-Level Controller DQN

Given that we have adopted a distributed ground zone approach with multiple intelligent agents and treated each ground zone as an individual agent, in this architecture, each ground zone needs to balance resource allocation and services admission decisions while collaborating and interacting with others. This involves finding a balance between cooperation and competition among multiple intelligent agents. Moreover, we recognize that the upper-level controller comprehensively considers the dynamic situations of each ground zone based on global information, thereby partially balancing the resource competition process among them and indirectly providing a global perspective for the lower-level controller. Therefore, we opt for a decentralized training approach for the deployment of lower-level controller agents. In this scheme, each ground zone’s agent is relatively independent in its training, making decisions based on local information. Here, we provide an overview of the training process for a single intelligent agent.
The lower-level controller primarily focuses on training the online DQN, and its principle is similar to training the online critic network in the upper-level controller. Similarly, based on the sample N ( s i , m d o w n , a i , m d o w n , s i + 1 , m d o w n , r i , m d o w n ) , we train the online DQN by minimizing the loss function:
L o s s ( μ ) = 1 N i N ( y i , m d o w m Q ( s i , m d o w n , a i , m d o w n | μ ) ) 2 ,
y i , m d o w m = r i , m d o w n + γ Q ( s i + 1 , m d o w n , a i + 1 , m d o w n | μ ) ,
a i + 1 , m d o w n = π ( s i + 1 , m d o w n | μ ) .
The parameter update of the target DQN still adopts the soft update method:
μ = τ μ + ( 1 τ ) μ .

4.3. Intelligent Hierarchical Admission Control Strategy

Considering the characteristics of satellite high-speed movement and the real-time variability of users’ needs and demands, we have adopted a strategy of online decision-making and offline learning for the training and decision-making of intelligent agents. This strategy aims to ensure that the agents can make adaptive decisions in real-time environments and continuously enhance their decision-making capabilities through offline learning.
During the online decision-making phase, the intelligent agents interact with the environment through the upper and lower-level controllers, accumulating experiences. This real-time interaction process enables the agents to promptly perceive changes in the satellite movement, the distribution of user services, and the actual utilization of resources. In the offline learning phase, we utilize these accumulated experiences to optimize the agents’ decision-making capabilities. By deeply learning and analyzing the experiential data, the agents can acquire insights into more intricate patterns and trends. The detailed strategy process is illustrated in Table 3.
Overall, the dual approach of online decision-making and offline learning allows the agents to balance the real-time responsiveness and learning capability. This strategy not only enables the agents to make timely decisions in dynamic environments but also continuously enhances their decision-making abilities through offline learning, thereby achieving a more intelligent and adaptive hierarchical admission control system.

5. Simulation Analysis

5.1. Simulation Parameter Settings

According to the relevant parameters of the Iridium constellation, a simulation environment is built by STK to obtain the satellite coverage time of each ground zone. Then, we establish a simulation system, and implement an intelligent hierarchical admission control strategy through MATLAB. Its simulation parameters are shown in Table 4.

5.2. Simulation Result Analysis

In the field of DRL, the convergence of algorithms is regarded as a key indicator for evaluating the performance and effectiveness of the algorithm. Especially when applied to LEO satellite communication systems, the convergence speed and stability of the reward function directly reflect the optimization level of the algorithm and its adaptability under different circumstances. Therefore, we observed the changes in the reward function during model training to obtain the optimal settings for model parameters. Furthermore, in simulation experiments, the proposed IHAC strategy is compared with DAQC and WOA in terms of channel user quantity, successful access probability, and drop call rate, demonstrating the applicability and superiority of the proposed strategy.
Figure 3 and Figure 4 depict the reward performance of the upper- and lower-level controllers during the training process, where the x-axis represents the number of training episodes and the y-axis represents the reward value. In DRL algorithms, the objective is to learn to select the optimal policy to maximize the cumulative reward. The reward discount factor can be used to balance the immediate reward at the current time step and the future rewards, with a larger discount factor focusing more on future rewards, allowing the controller to consider the long-term impact of actions and promoting algorithm convergence. Therefore, in our experiments, we set the reward discount factors as γ = { 0.9 , 0.5 , 0.1 } . From the figures, it can be observed that the reward values tend to stabilize during the algorithm iterations. However, when γ = 0.9 , the reward function stabilizes more rapidly and remains stable without significant fluctuations over a longer period. This indicates that when γ = 0.9 , the upper- and lower-level controllers exhibit a higher adaptability to different communication scenarios and service demands in the LEO satellite communication system.
In this paper, simulation verification is conducted using a time division approach, where a single resource allocation time window is divided into 10 units of simulated time. Figure 5 presents the changing trend of the number of user services on the channel as the simulation progresses. From the figure, it is evident that the intelligent hierarchical admission control strategy interacts continuously with beam resources and user service states. This interaction, constrained by the reward function, maximizes the utilization of beam resources. By continuously interacting with beam resources and user service states through hierarchical admission control, the maximum utilization of beam resources is achieved, thereby avoiding resource idleness.
The proposed strategy in this paper shows a significant improvement in channel utilization after sufficient training. It not only avoids resource wastage but also effectively meets the access demands of user services. Compared to DAQC and WOA, the proposed strategy achieves the highest received service count of 280 and 320, respectively, while the accepted service count is increased by 22.95% and 21.19%, respectively. The results highlight that the proposed intelligent strategy, after training, significantly enhances the utilization of channels, not only preventing resource wastage but also efficiently satisfying the access requirements of user services.
The rate of successful service access is a critical parameter that reflects user service satisfaction and is one of the core goals of satellite networks. A high rate of successful service access implies that services are more likely to be effectively provided. Especially in scenarios with numerous service demands, terminal users need to compete for limited beam resources. Therefore, the level of this indicator significantly impacts user experience. As shown in Figure 6, as the number of services increases, the rate of successful service access gradually decreases. This is due to network congestion that might prevent accommodating all service access requests. However, the strategy proposed in this paper maintains a relatively high rate of successful service access across various service quantity scenarios. Compared to DAQC and WOA, the average success access rate is increased by 12.12% and 19.05%, respectively. This demonstrates the strategy’s remarkable adaptability in highly competitive environments.
Compared to services that have not been accessed yet, services that are interrupted after accessing are more challenging to accept. This not only means that the resources allocated to the service are released but that this also requires reapplication and resource competition to continue the service. As shown in Figure 7, the service drop rate changes based on different service quantities. The graph clearly indicates that the proposed strategy maintains a lower drop rate across various service quantity scenarios. This further underscores the strategy’s efficiency in resource allocation and admission decisions, reducing service interruptions and dropped calls, thus enhancing user experience and satisfaction.
The fairness of resource allocation is defined by the ratio of allocated resources to requested resources, which reflects the satisfaction of ground zone services. Achieving fair resource allocation can reduce unnecessary resource competition. An analysis of Figure 8 indicates that in the initial stages of the strategy, due to the relatively low number of services initially connected to the satellite and the relatively abundant system resources, the fairness of resource allocation is relatively high. However, as the number of requested communication services gradually increases, differences in dynamic priorities among ground zones and within each zone’s services become more apparent, leading to a gradual decrease in the fairness of resource scheduling. During the process where the number of services reaches 1000, the strategy proposed in this paper reaches and maintains stable resource fairness at the fastest pace. It outperforms DAQC and WOA in terms of resource fairness. This demonstrates the strategy’s exceptional adaptability when facing dynamic and diverse priorities. It maintains high fairness even as the number of services increases, showcasing its ability to effectively handle varying priority scenarios.
Finally, we show the performance comparison between IHAC and baseline algorithms with average metric data, as shown in Table 5. In addition, we analyze the time and space complexity of IHAC. As we mentioned before, the ground zone and user service status characteristics in each zone are reported to the upper and lower controllers, respectively, in order to train and run the DDPG and DQN in the controller. Since all training and running tasks are performed by the controller, most of the computation and storage cost is incurred in the controller.
In the controller, the computation related to DRL consumes most of the resources, and we focus on analyzing its time and space complexity. During the inference and operation of DDPG and DQN, their computational overhead mainly depends on the number of nodes in the input layer. Specifically, if M represents the number of ground zones, K represents the characteristic dimension of each ground zone, N represents all user services within the satellite coverage, and L represents the characteristic dimension of user services, then the input layer of DDPG consists of MK units, and the input layer of DQN consists of NL units. Then the time complexity can be expressed as O ( | M K | 2 ) + O ( | N L | 2 ) . Since DAQC also uses the Markov decision process to model the problem, which is similar to the DRL modeling process, we estimate that its computational complexity is comparable to IHAC. However, for swarm intelligence heuristic algorithms such as WOA, the computational complexity is related to the dimension of search space D and the time complexity f of the fitness function. That is, it can be roughly expressed as O ( | M | 2 ( D + f ) ) + O ( | N | 2 ( D + f ) ) . In addition, the space complexity of the upper and lower controllers can be expressed as O ( M K ) + O ( N L ) . However, in the training process, we only need to use a set of samples to carry out the forward propagation of DDPG and DQN. So, the actual space complexity is much less than O ( M K ) + O ( N L ) .
We can find that DL techniques require more computation and storage costs compared to traditional strategies. However, the IHAC we consider can not only achieve the adaptive ability to the dynamic environment through the feature extraction of massive data, but also achieve the dual goals of efficient utilization of beam resources and service fairness.

6. Conclusions

This paper proposes an intelligent hierarchical admission control (IHAC) strategy based on DRL for LEO satellites. The strategy addresses the resource allocation and admission control challenges in the field of multi-beam satellite communications, specifically focusing on intelligent admission control for ground zone services within a multi-beam coverage area. The strategy employs DDPG and DQN algorithms to create a hierarchical control strategy that coordinates global resource scheduling with local service admission. The integration of user service models and adaptive dynamic priority models ensures a balance between resource utilization and fairness.
By introducing well-designed state feature inputs and reward mechanisms, the strategy guides decision-making to achieve efficient resource utilization and fairness. In summary, the IHAC strategy proposed in this paper fills the gap in the field of satellite multi-beam communications by providing a comprehensive and effective hierarchical admission control strategy. Leveraging the power of DRL, this strategy can cater to diverse service demands, ensuring both efficient resource utilization and service fairness. It presents a solution for the optimization and advancement of multi-beam satellite communication systems.

Author Contributions

Conceptualization, D.W.; data curation, C.G.; formal analysis, D.W.; investigation, D.W.; methodology, D.W.; project administration, L.Y.; resources, C.G.; software, C.G.; supervision, L.Y.; validation, D.W.; visualization, C.G.; writing—original draft, C.G.; writing—review and editing, D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly financially supported through grants from the National Natural Science Foundation of China (U21B2003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xia, S.; Jiang, Q.; Zou, C.; Li, G. Beam Coverage Comparison of LEO Satellite Systems Based on User Diversification. IEEE Access 2019, 7, 181656–181667. [Google Scholar] [CrossRef]
  2. Lu, K.; Liu, H.; Zeng, L.; Wang, J.Y.; Zhang, Z.S.; An, J.P. Applications and prospects of artificial intelligence in covert satellite communication: A review. Sci. China Inf. Sci. 2023, 66, 121301. [Google Scholar] [CrossRef]
  3. Leng, T.; Wang, Y.; Hu, D.; Cui, G.; Wang, W. User-level scheduling and resource allocation for multi-beam satellite systems with full frequency reuse. China Commun. 2022, 19, 179–192. [Google Scholar] [CrossRef]
  4. Kisseleff, S.; Lagunas, E.; Abdu, T.S.; Chatzinotas, S.; Ottersten, B. Radio Resource Management Techniques for Multibeam Satellite Systems. IEEE Commun. Lett. 2021, 25, 2448–2452. [Google Scholar] [CrossRef]
  5. Cocco, G.; Cola, T.D.; Angelone, M.; Katona, Z.; Erl, S. Radio Resource Management Optimization of Flexible Satellite Payloads for DVB-S2 Systems. IEEE Trans. Broadcast. 2018, 64, 266–280. [Google Scholar] [CrossRef]
  6. Lin, Z.; Lin, M.; Champagne, B.; Zhu, W.P.; Al-Dhahir, N. Secrecy-energy efficient hybrid beamforming for satellite-terrestrial integrated networks. IEEE Trans. Commun. 2021, 69, 6345–6360. [Google Scholar] [CrossRef]
  7. Lin, Z.; Niu, H.; An, K.; Wang, Y.; Zheng, G.; Chatzinotas, S.; Hu, Y. Refracting RIS-aided hybrid satellite-terrestrial relay networks: Joint beamforming design and optimization. IEEE Tran. Aerosp. Electron. Syst. 2022, 58, 3717–3724. [Google Scholar] [CrossRef]
  8. Lin, Z.; An, K.; Niu, H.; Hu, Y.; Chatzinotas, S.; Zheng, G.; Wang, J. SLNR-based secure energy efficient beamforming in multibeam satellite systems. IEEE Trans. Aerosp. Electron. Syst. 2022, 59, 2085–2088. [Google Scholar] [CrossRef]
  9. Deng, Z.; Yu, X.; Lin, W.; Wang, K.; Liu, H.; Gu, L.; Liu, Y.; Ma, X. A Multi-Beam Satellite Cooperative Transmission Scheme Based on Resources Optimization and Packets Segmentation. Electronics 2021, 10, 2841. [Google Scholar] [CrossRef]
  10. Zhang, X.; Zhang, B.; Guo, D.; An, K.; Qi, S.; Wu, G. Potential Game-Based Radio Resource Allocation in Uplink Multibeam Satellite IoT Network. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 4269–4279. [Google Scholar] [CrossRef]
  11. Takahashi, M.; Kawamoto, Y.; Kato, N.; Miura, A.; Toyoshima, M. Adaptive Power Resource Allocation with Multi-Beam Directivity Control in High-Throughput Satellite Communication Systems. IEEE Wirel. Commun. Lett. 2019, 8, 1248–1251. [Google Scholar] [CrossRef]
  12. Jia, M.; Zhang, X.; Gu, X.; Guo, Q.; Li, Y.; Lin, P. Interbeam Interference Constrained Resource Allocation for Shared Spectrum Multibeam Satellite Communication Systems. IEEE Internet Things 2019, 6, 6052–6059. [Google Scholar] [CrossRef]
  13. Wang, Y.; Bian, D.; Hu, J.; Tang, J.; Wang, C. A Flexible Resource Allocation Algorithm in Full Bandwidth Beam Hopping Satellite Systems. In Proceedings of the 2019 IEEE 3rd Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 11–13 October 2019. [Google Scholar]
  14. Cai, B.; Zhang, Q.; Ge, J.; Xie, W. Resource Allocation for Cognitive LEO Satellite Systems: Facilitating IoT Communications. Sensors 2023, 23, 3875. [Google Scholar] [CrossRef] [PubMed]
  15. Kaloxylos, A.; Gavras, A.; Camps, D.; Ghoraishi, M. AI and ML–Enablers for beyond 5G Networks; Centres de Recerca de Catalunya (CERCA): Catalonia, Spain, 2021. [Google Scholar]
  16. Homssi, B.A.; Dakic, K.; Wang, K.; Allen, B.; Kandeepan, S.; AI-Hourani, A.; Saad, W. Artificial intelligence techniques for next-generation mega satellite networks. arXiv 2022, arXiv:2207.00414. [Google Scholar]
  17. Trakadas, P.; Masip-Bruin, X.; Facca, F.M.; Spantideas, S.T.; Giannopoulos, A.E.; Kapsalis, N.C.; Martins, R.; Bosani, E.; Ramon, J.; Prats, R.G.; et al. A Reference Architecture for Cloud–Edge Meta-Operating Systems Enabling Cross-Domain, Data-Intensive, ML-Assisted Applications: Architectural Overview and Key Concepts. Sensors 2022, 22, 9003. [Google Scholar] [CrossRef]
  18. Lyu, Y.; Liu, Z.; Fan, R.; Zhan, C.; Hu, H.; An, J. Optimal computation offloading in collaborative leo-iot enabled mec: A multi-agent deep reinforcement learning approach. IEEE Trans. Green Commun. Netw. 2022, 7, 996–1011. [Google Scholar] [CrossRef]
  19. Karamplias, T.; Spantideas, S.T.; Giannopoulos, A.E.; Gkoniset, P.; Kapsalisal, N.; Trakadas, P. Towards closed-loop automation in 5g open ran: Coupling an open-source simulator with xapps. In Proceedings of the 2022 Joint European Conference on Networks and Communications & 6G Summit (EuCNC/6G Summit), Grenoble, France, 7–10 June 2022. [Google Scholar]
  20. Liao, X.; Hu, X.; Liu, Z.; Ma, S.; Ghannouchi, F.M. Distributed intelligence: A verification for multi-agent DRL-based multibeam satellite resource allocation. IEEE Commun. Lett. 2020, 24, 2785–2789. [Google Scholar] [CrossRef]
  21. Chan, S.; Jo, G.; Kim, S.; Oh, D.; Ku, B.J. Dynamic power and bandwidth allocation for DVB-based LEO satellite systems. ETRI J. 2022, 44, 955–965. [Google Scholar] [CrossRef]
  22. Hu, X.; Liao, X.; Liu, Z.; Liu, S.; Ding, X.; Helaoui, M.; Wang, W.D.; Ghannouchi, F.M. Multi-Agent Deep Reinforcement Learning-Based Flexible Satellite Payload for Mobile Terminals. IEEE Trans. Veh. Technol. 2020, 69, 9849–9865. [Google Scholar] [CrossRef]
  23. He, Y.; Sheng, B.; Yin, H.; Yan, D.; Zhang, Y. Multi-objective deep reinforcement learning based time-frequency resource allocation for multi-beam satellite communications. China Commun. 2022, 19, 77–91. [Google Scholar] [CrossRef]
  24. Huang, Y.; Wu, S.F.; Zeng, Z.K.; Kang, Z.Y.; Mu, Z.C.; Huang, H. Sequential dynamic resource allocation in multi-beam satellite systems: A learning-based optimization method. Chin. J. Aeronaut. 2023, 36, 288–301. [Google Scholar] [CrossRef]
  25. Dai, J.; Geng, N.; Xie, X. Dynamic admission quota control with controllable and uncontrollable demands and random service time. IEEE Trans. Automat. Contr. 2020, 66, 2925–2932. [Google Scholar] [CrossRef]
  26. Goudos, S.K.; Diamantoulakis, P.D.; Boursianis, A.D.; Sarigiannidis, P.; Psannis, K.E.; Matin, M.A.; Wan, S.; Karagiannidis, G.K. Joint QoS Aware Admission Control and Power Allocation in NOMA Downlink Networks. IEEE Access 2023, 11, 30873–30887. [Google Scholar] [CrossRef]
  27. Li, Z.; Xie, Z.; Liang, X. Dynamic channel reservation strategy based on DQN algorithm for multi-service LEO satellite communication system. IEEE Wirel. Commun. Lett. 2020, 10, 770–774. [Google Scholar] [CrossRef]
  28. Bakri, S.; Brik, B.; Ksentini, A. On using reinforcement learning for network slice admission control in 5G: Offline vs. online. Int. J. Commun. Syst. 2021, 34, e4757. [Google Scholar] [CrossRef]
  29. Wu, Z.; Ishigaki, G.; Gour, R.; Jue, J.P. A reinforcement learning-based admission control strategy for elastic network slices. In Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM), Madrid, Spain, 7–11 December 2021. [Google Scholar]
  30. Choi, H.; Pack, S. Cooperative Downloading for LEO Satellite Networks: A DRL-Based Approach. Sensors 2022, 22, 6853. [Google Scholar] [CrossRef]
Figure 1. Multi-beam satellite communication system model.
Figure 1. Multi-beam satellite communication system model.
Sensors 23 08470 g001
Figure 2. Intelligent hierarchical admission control framework.
Figure 2. Intelligent hierarchical admission control framework.
Sensors 23 08470 g002
Figure 3. Reward of the upper-level controller under different reward discount factors.
Figure 3. Reward of the upper-level controller under different reward discount factors.
Sensors 23 08470 g003
Figure 4. Reward of the lower-level controller under different reward discount factors.
Figure 4. Reward of the lower-level controller under different reward discount factors.
Sensors 23 08470 g004
Figure 5. Number of user services on the channel.
Figure 5. Number of user services on the channel.
Sensors 23 08470 g005
Figure 6. Rate of successful service access.
Figure 6. Rate of successful service access.
Sensors 23 08470 g006
Figure 7. Drop call rate of user services.
Figure 7. Drop call rate of user services.
Sensors 23 08470 g007
Figure 8. Fairness in resource allocation.
Figure 8. Fairness in resource allocation.
Sensors 23 08470 g008
Table 1. User service model parameters.
Table 1. User service model parameters.
NotationDefinition
B t o t a l The total bandwidth of the multi-beam satellite
P t o t a l The total power of the multi-beam satellite
U m , n The n th user service model in the m th ground zone
B m , n Requested bandwidth of U m , n
P m , n Requested power of U m , n
C m , n Requested capacity of U m , n
V m , n Initial priority of user services U m , n
T s m , n The arrival time of the user’s service U m , n
S m , n The time required to complete the U m , n
t The current time
S T , E T The start and end time of the current resource allocation time window
D P m , n Dynamic priority model for U m , n
Table 2. DRL formulation parameters.
Table 2. DRL formulation parameters.
NotationDefinition
N m The number of user services requests in the m th ground zone
P r m The total dynamic priority of user services in the m th ground zone
W t m The total waiting time of user services in the m th ground zone
B m , P m , C m The total requested bandwidth, power, and capacity of user services in the m th ground zone
B ^ m , n , P ^ m , n , C ^ m , n The ratio of the requested bandwidth, power, and capacity of the n th user service to the total requested resources in the m th ground zone
S P R m The total dynamic priority of successfully admitted services in the m th ground zone
S P R a v e The average total dynamic priority of successfully admitted services across all ground zones
DWTmThe total waiting time of services not admitted in the m th ground zone
D W T a v e The average total waiting time of services not admitted across all ground zones
S W T m The total waiting time of successfully admitted services in the m th ground zone
R P m The reward or penalty for the resource utilization rate of the ground area admission control strategy
R U m The ratio of resources allocated to user services within the m th ground zone
Table 3. Intelligent hierarchical admission control strategy.
Table 3. Intelligent hierarchical admission control strategy.
1:
Initializing the neural network parameters θ , ω , and μ in the upper and lower-level controllers. Let θ = θ , ω = ω , and μ = μ .
2:
Initializing the experience replay buffers, training episode T E in the upper and lower-level controllers.
3:
Initializing the sampling episode S E with 0.
4:
Initializing the state features s t u p = ( s t , 1 u p , s t , 2 u p , , s t , m u p , , s t , M u p , s t , B u p , s t , P u p ) .
5:
For each sampling episode S E , do:
6:
The online actor network of the upper-level controller computes the scheduling strategy of beam resources for each ground zone based on the input state s t u p , i.e., a t u p = π ( s t u p | θ ) . Subsequently, the scheduling strategy a t , m u p is transmitted to the corresponding lower-level controller.
7:
The lower-level controllers deployed in each ground zone combine the local service information with the upper-level strategy a t , m u p to make decisions. The online DQN then makes admission control decisions for each user service within the ground zone based on the input state, i.e., a t , m d o w n = π ( s t , m d o w n | μ ) .
8:
The upper-level beam resource scheduling strategy and the lower-level user service admission control strategy are executed, resulting in states s t u p and s t , m d o w n , as well as rewards r t u p and r t , m d o w n . These components form experience samples ( s t u p , a t u p , s t + 1 u p , r t u p ) and ( s t , m d o w n , a t , m d o w n , s t + 1 , m d o w n , r t , m d o w n ) , which are stored in the upper-level and lower-level experience replay pools, respectively.
9:
Let s t u p = s t + 1 u p and s t , m d o w n = s t + 1 , m d o w n .
10:
If S E > T E :
11:
Samples N ( s i u p , a i u p , s i + 1 u p , r i u p ) and N ( s i , m d o w n , a i , m d o w n , s i + 1 , m d o w n , r i , m d o w n ) are obtained through random sampling, and then the neural networks of the upper-level and lower-level controllers are trained and updated using Equations (15)–(23).
12:
End for.
Table 4. Parameter setting.
Table 4. Parameter setting.
ParameterValue
Satellite total power, P t o t a l 30 dBW
Satellite total bandwidth, B t o t a l 500 MHz
Satellite height600 km
Number of services within satellite coverage area2000
Number of ground zones40
Simulation time unit length, T s 50 ms
Initial priority of user services, V i 1, 2, 3, 4, 5
Initial required capacity of services, C i [2, 4, 8, 12, 16, 24] × 300 bps
Discount factor for rewards, γ 0.9
Soft update learning rate, τ 0.001
Training episode, T E 300
Experience replay buffers5000
Sampling batch size128
Table 5. Performance comparison of IHAC with baseline algorithms.
Table 5. Performance comparison of IHAC with baseline algorithms.
MetricsIHACDAQCWOA
Average number of user services on the channel1486.812951298
Average call drop rate of user services1.2561.4881.623
Average probability of successful service access0.8810.7850.739
Average fairness in resource allocation20.66817.52617.168
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, D.; Guo, C.; Yang, L. Intelligent Hierarchical Admission Control for Low-Earth Orbit Satellites Based on Deep Reinforcement Learning. Sensors 2023, 23, 8470. https://doi.org/10.3390/s23208470

AMA Style

Wei D, Guo C, Yang L. Intelligent Hierarchical Admission Control for Low-Earth Orbit Satellites Based on Deep Reinforcement Learning. Sensors. 2023; 23(20):8470. https://doi.org/10.3390/s23208470

Chicago/Turabian Style

Wei, Debin, Chuanqi Guo, and Li Yang. 2023. "Intelligent Hierarchical Admission Control for Low-Earth Orbit Satellites Based on Deep Reinforcement Learning" Sensors 23, no. 20: 8470. https://doi.org/10.3390/s23208470

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop