Next Article in Journal
Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears
Next Article in Special Issue
An Effective Gaze-Based Authentication Method with the Spatiotemporal Feature of Eye Movement
Previous Article in Journal
Non-Contact Paper Thickness and Quality Monitoring Based on Mid-Infrared Optical Coherence Tomography and THz Time Domain Spectroscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Deep-Learning-Based Resource Allocation for Time-Sensitive Device-to-Device Networks

1
State Grid Key Laboratory of Power Industrial Chip Design and Analysis Technology, Beijing Smart-Chip Microelectronics Technology Co., Ltd., Beijing 102299, China
2
College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou 310058, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1551; https://doi.org/10.3390/s22041551
Submission received: 28 December 2021 / Revised: 12 February 2022 / Accepted: 14 February 2022 / Published: 17 February 2022
(This article belongs to the Special Issue Trustworthy Sensing with Human-and-Environment-in-the-Loop)

Abstract

:
Ultra-reliable and low-latency communication (URLLC) is considered as one of the major use cases in 5G networks to support the emerging mission-critical applications. One of the possible tools to achieve URLLC is the device-to-device (D2D) network. Due to the physical proximity of communicating devices, D2D networks can significantly improve the latency and reliability performance of wireless communication. However, the resource management of D2D networks is usually a non-convex combinatorial problem that is difficult to solve. Traditional methods usually optimize the resource allocation in an iterative way, which leads to high computational complexity. In this paper, we investigate the resource allocation problem in the time-sensitive D2D network where the latency and reliability performance is modeled by the achievable rate in the short blocklength regime. We first design a game theory-based algorithm as the baseline. Then, we propose a deep learning (DL)-based resource management framework using deep neural network (DNN). The simulation results show that the proposed DL-based method achieves almost the same performance as the baseline algorithm, while it is more time-efficient due to the end-to-end structure.

1. Introduction

The focus of the fifth-generation (5G) wireless communication networks is to provide reliable services for various applications with the design objectives of high throughput, reduced end-to-end latency, and massive device connectivity [1]. To support the emerging mission-critical applications (such as vehicle-to-vehicle communications, industrial automation, and virtual reality) that are sensitive to time delay and transmission reliability [2], it is necessary for wireless communication systems to achieve high quality of service (QoS). Therefore, ultra-reliable and low-latency communication (URLLC), which intrinsically takes reliability and latency into consideration during network design, is considered as one of the major use cases in 5G networks.
Aiming at achieving an ultra-low packet error rate (e.g., 10 5 ) and an ultra-low latency (e.g., 1 ms) [3,4], URLLC brings new challenges to the network design. Due to the inherent instability of wireless channel, it is usually difficult to achieve low latency and high reliability simultaneously. The reliability can be improved by allocating more radio resources, such as spectrum, re-transmissions, etc., which leads to the increase of transmission latency and the degradation of throughput performance [5]. Therefore, it is necessary to develop novel system architectures and network layouts to satisfy URLLC requirements by emphasizing high reliability and low latency simultaneously.
To support the increasingly heavy local traffic in wireless services, a device-to-device (D2D) communication enabled cellular network has been widely studied and is regarded as one of the key components in 5G networks. Due to the physical proximity of communicating devices, D2D communications can significantly improve the spectrum and energy efficiency [6]. In addition, the direct communication link between D2D pairs, as the most important feature of D2D communications, can also reduce the end-to-end latency of data transmission, which makes D2D communications a powerful tool to achieve URLLC in 5G networks. In general, a D2D network consists of several D2D pairs that reuse the spectrum of cellular networks. There are mainly two reusing modes in the existing literature, namely underlay mode and overlay mode [7]. In the underlay mode, D2D pairs transmit data on the spectrum allocated to cellular users (CUs), resulting in the interference between D2D pairs and CUs. To improve the system performance of the underlay mode, some methods have been developed to alleviate the co-channel interference, including mode selection [8] and resource allocation [9]. In the overlay mode, a dedicated spectrum is allocated to D2D pairs to allow interference-free D2D transmissions. In this paper, we focus on the overlay mode to avoid the performance degradation of CUs. Although D2D communications can be conducted without interference in the overlay mode, the reserved spectrum is usually rather limited, which makes it infeasible to allocate spectrum orthogonally for D2D pairs. Multiple D2D pairs may transmit data on the same channel with severe co-channel interference, resulting in the degradation of system performance [10]. Therefore, it is necessary to determine the proper channel and power allocation for D2D pairs to guarantee the system performance, including throughput, latency, and reliability.

1.1. Related Works

The radio resource allocation problem has been widely studied to improve the system performance of D2D networks. In [11], mode selection and channel-power allocation have been both considered to improve the throughput performance of D2D networks. In [12], a matching algorithm based on game theory has been proposed to improve the energy efficiency of D2D transmission. In [13], the sum-rate is maximized under different rate requirements of D2D pairs. Although several system metrics including throughput and energy efficiency have been widely investigated in the existing works, there are not sufficient studies on the URLLC performance of D2D networks. In [14], a power control policy has been proposed to achieve the maximum average rate in the finite block-length regime with a reliability constraint. However, the system model in this work is too simple, where the system only consists of a D2D pair and a cellular user. Leveraging the game theory, the successful transmission probability of an overlay D2D network is maximized by achieving Nash equilibrium [15]. Nevertheless, the power control strategy is not investigated, and the throughput performance of the D2D network is not guaranteed. In [16], the authors have proposed a joint reservation and contention-based channel allocation strategy to minimize the sum packet error probability of D2D networks. In [17], the authors have proposed a D2D-based protocol to improve the reliability of downlink transmissions. In [18], a two-timescale resource allocation framework has been proposed, where the user pairing is conducted at the long timescale and the transmission time is determined at the short timescale.
In recent years, deep learning (DL) has been successfully applied to a variety of fields, such as computer vision, natural language processing, and speech recognition. In the communication systems, a variety of DL-based methods have also been used in resource allocation [19], user association [20], channel estimation [21], and physical layer processing [22]. In [23], a resource allocation strategy based on deep neural network (DNN) has been designed to increase the overall spectrum efficiency of an underlay D2D network. In [24], the channel selection and transmit power of the D2D pairs are jointly optimized to achieve a higher weighted sum rate by using deep reinforcement learning. However, the DL-based resource management method for URLLC-oriented D2D networks has not been investigated in the existing literature.

1.2. Motivation and Contribution

Although the existing methods can improve the URLLC performance of D2D networks by properly allocating the radio resources, the throughput performance of D2D networks at the finite blocklength has not been investigated. Moreover, the high computational complexity brought by the iterative structure may lead to the degradation of latency performance, which motivates us to develop a time-efficient resource management approach for D2D networks based on DL. To this end, in this paper, we propose a resource allocation method for overlay D2D networks based on DL. The main contributions of this paper are summarized as follows.
  • Instead of Shannon rate, we adopt the short packet coding rate to more accurately capture the rate loss in the finite blocklength regime. The reliability of D2D transmissions is guaranteed by choosing a proper coding rate lower than the achievable rate.
  • We propose an iterative channel selection and power allocation algorithm based on game theory. The sum utility function of all D2D pairs is maximized by alternately updating the channel selection and power allocation of each D2D pair.
  • To improve the time efficiency of the resource allocation procedure, we propose two DNN-based methods to solve the sum rate maximization problems with and without the minimum rate constraint, respectively. The network structure and output are properly designed to improve the learning ability of the network.
The rest of this paper is organized as follows. Section 2 introduces the system model and formulates the sum rate maximization problem for D2D networks. In Section 3, the game theory-based algorithm is proposed. In Section 4, two DNN-based methods are proposed for resource allocation in D2D networks. In Section 5, simulation results are provided to demonstrate the effectiveness of the proposed algorithms. Finally, the article is concluded in Section 6.

2. System Model and Problem Formulation

2.1. System Model and Transmission Model

In this paper, we consider an overlay multi-channel D2D network, as depicted in Figure 1. In particular, N D2D pairs are arbitrarily distributed and share K orthogonal wireless channels, where N is typically larger than K. Each D2D pair consists of one transmitter and one receiver. Let N = { 1 , , N } denote the set of D2D pairs and K = { 1 , , K } denote the set of channels. We consider a slotted OFDM transmission scheme where the length of a time slot is T and the bandwidth of a channel is B. Due to the strict latency requirement of the D2D pairs, T is assumed to be rather small to shorten the transmission latency.
Let h n , m k = α n , m g n , m k denote the channel gain of the k-th channel between the transmitter of D2D pair n and the receiver of D2D pair m, where α n , m and g n , m k are the large-scale and small-scale channel fading, respectively. The large-scale channel fading is composed of path loss and shadowing, as α n , m = G ψ n , m d n , m ϕ , where G is the pathloss constant, ψ n , m is the log-normal shadowing random component, d n , m is the distance between the transmitter of D2D pair n and the receiver of D2D pair m, and ϕ represents the path loss exponent. The small-scale channel gain g n , m k is assumed to be Rayleigh fading, which is independent and identically distributed (i.i.d) with unit mean for different n , m , and k. It is noted that h n , m k represents the local D2D channel when n = m and represents the interference D2D channel when n m . In practical communication systems, the real-time channel state information (CSI) is hard to obtain due to the dynamic wireless environment. To tackle this problem, the accurate CSI could be substituted with the position information of the D2D pairs, which is much easier to obtain. We assume that the position of the D2D pairs does not change in a certain time interval. Therefore, the average large-scale channel gain can be calculated as G d ϕ . By using the large-scale channel gain as the input of the proposed methods, the resource allocation strategy could be obtained in the absence of accurate CSI.
It is assumed that each D2D pair can only occupy a single resource block ( R B ) for data transmission, which can relieve the channel congestion and reduce the computational complexity at the receiver. Let ρ n k be the channel indicator of D2D pair n; i.e., ρ n k = 1 indicates that D2D pair n is transmitting data through channel k and ρ n k = 0 otherwise. Then, the SINR of D2D pair n can be expressed as
γ n = k = 1 K ρ n k p n h n , n k k = 1 K l = 1 , l n N ρ n k ρ l k p l k h l , n k + N 0 B ,
where p n is the transmit power of the D2D pair n and N 0 is the power spectral density of noise.
Due to the strict latency constraint of URLLC services, the slot duration of D2D transmission is very short. Therefore, the traditional Shannon channel capacity that assumes the code length to be infinite does not hold in this case. With the bandwidth B and slot duration T, the maximal achievable rate of D2D pair n in the finite blocklength regime is accurately approximated by [25]
R n = B log 2 ( 1 + γ n ) B V n T Q 1 ( ϵ ) ln 2 ,
where Q 1 ( · ) is the inverse of the Gaussian Q-function, ϵ is the packet error probability, and V n is the channel dispersion given by [25]
V n = 1 1 + γ n 2 1 .
According to [26], the approximation in (3) is quite accurate in the high SNR regime.

2.2. Problem Formulation

In the D2D network, each D2D pair chooses a single channel and a certain transmit power according to the global channel information. We first formulate the achievable rate maximization problem without minimum rate constraint as
P 1 : max ρ n k , P n n = 1 N R n ,
s . t . k = 1 K ρ n k = 1 , n N , k K ,
ρ n k { 0 , 1 } , n N , k K ,
p n P max , n N ,
p n 0 , n N .
The objective of problem P 1 is to maximize the sum finite blocklength rate of all D2D pairs, where constraints (4b), (4c) ensure that each D2D pair can only choose a single channel and (4d), (4e) are the transmit power constraints.
In some cases, the D2D pairs may have a minimum rate requirement on the communication links to achieve better QoS. To address this requirement, the achievable rate maximization problem with a minimum rate constraint can be formulated as
P 2 : max ρ n k , P n n = 1 N R n ,
s . t . R n R n min , n N , ( 4 b ) ( 4 e ) .
Problems P 1 and P 2 are both mixed integer nonlinear programming (MINLP) problems with non-convex objective functions, which are hard to solve. Although some heuristic algorithms can be applied to find the suboptimal solution, their iterative structures with high computational complexity may lead to the degradation of latency performance. The classic branch-and-bound (BnB) algorithm can find global optimal solutions, but its complexity scales exponentially with the problem size. In this problem, the complexity of the BnB algorithm grows significantly as the user number increases, which is challenging for real-time implementation. On the other hand, the recent successful applications of machine learning techniques demonstrate that DL has the ability to learn various patterns and approximate arbitrary function mappings [27]. Moreover, the time complexity of the DL method is usually much lower than that of the classic methods, which makes DL a competitive tool when solving communication problems. In fact, DL has been successfully applied to the channel selection and power control in D2D networks [23,24]. Therefore, we adopt the powerful DL method to solve the formulated MINLP problem by directly mapping the input data to the channel selection and power allocation strategy.

3. Game Theory-Based Resource Allocation

Since the complexity of the global optimal algorithm scales exponentially with the problem size, it is necessary to propose a suboptimal solution with low computational complexity. In this section, we propose a channel selection and power allocation algorithm based on spatial adaptive play (SAP), which is a high-profile algorithm in potential games because it can achieve the optimal Nash equilibrium. In fact, SAP has been successfully applied to the resource allocation problem in wireless communications [15], which motivates us to utilize SAP to solve the formulated problems. In SAP, one player is randomly chosen to perform learning in each iteration. The sum utility function of all D2D pairs is maximized by alternately updating the channel selection and power allocation of each D2D pair. To reduce the computational complexity, the transmit power is discretized into L discrete levels, that is, p n P = P 1 , P 2 , P L . For simplicity, let P = p 1 , p 2 , , p N denote the power vector that collects the transmit power of all D2D pairs. Let Φ denote the N × K -sized channel selection matrix where Φ n , k = ρ n k . Let H denote the N × N × K -sized channel matrix where H n , m , k = h n , m k .
The proposed algorithm for problems P 1 and P 2 is shown in Algorithm 1. In step 2, μ is the exploration probability, which represents the willingness of exploration when the D2D pairs choose the channel and power level in step 6. When μ decreases, the D2D pairs are less willing to explore and tend to choose the channel and power level with the highest utility function. t is the iteration index, w is the decay rate that controls the decreasing speed of μ , and τ is the aborting threshold. U t is the value of the utility function in the t-th iteration, which is different for problems P 1 and P 2 . To be specific, for problem P 1 , U = n = 1 N R n is the sum rate of D2D pairs. For problem P 2 , U = n = 1 N R n η max ( R n m i n R n , 0 ) is the penalized sum rate of D2D pairs, where η > 0 is a weight constant to ensure that the minimum rate constraint is satisfied. As η increases, the minimum rate constraint is better guaranteed, while the sum rate decreases because the interference becomes more severe. Therefore, the rate requirement of the users with bad channel condition could be satisfied by adopting a large η , although the sum rate performance might be sacrificed in this case.
Algorithm 1 SAP-Based Channel Selection and Power Allocation Algorithm
1:
Input: H , N , K , B , T , ϵ , σ 2 .
2:
Initialization: Set t = 0 , μ = 1 , w < 1 , τ > 0 , and U 0 = 0 . Initialize the channel selection matrix Φ and power matrix P.
3:
Repeat:
4:
• Set t = t + 1 .
5:
• Randomly choose a D2D pair n, calculate the sum rate on all possible channel selection ρ n , k , k K and power allocation p n . Assume that the utility function is maximized as U t when D2D pair n chooses channel k 0 with power level l 0 .
6:
• With probability 1 μ : Set ρ n k 0 = 1 , ρ n k = 0 , k k 0 , and p n = P l 0 . With probability μ : Set ρ n k 1 = 1 , ρ n k = 0 , k k 1 , and p n = P l 1 , where k 1 and l 1 are randomly picked from 1 , , K and 1 , , L , respectively.
7:
• Set μ = w μ .
8:
Until | U t U t 1 | < τ .
9:
Output: Φ and P.
In each iteration, the algorithm randomly chooses a D2D pair and finds the optimal channel selection and power allocation strategy through exhaustive search. In this way, the D2D pairs take turns to update the strategy to increase the sum rate step by step, which guarantees the convergence of the algorithm. On the other hand, the D2D pairs explore the possible actions with probability μ in each iteration. This can help the algorithm jump out of the local minimum and find better solutions. It is noted that the exploration probability keeps decreasing as the number of iterations increases, which means that the algorithm gradually stops exploring and finally converges to the solution.
After a number of iterations, the channel selection and power allocation of all D2D pairs will converge to a Nash equilibrium, which means that no D2D pair can achieve a higher value of the utility function by changing its channel selection and transmit power. The channel selection matrix Φ and power matrix P at this point are taken as the final output of the algorithm. It is noted that it takes a number of iterations for the SAP-based channel selection and power allocation algorithm to converge. The non-ignorable computational overhead may lead to the degradation of latency performance. Therefore, it is necessary to develop a time-efficient DL-based method to solve the channel selection and power allocation problem.

4. Deep Learning Based Resource Allocation

In this section, we develop a DNN-based channel selection and power allocation method for problems P 1 and P 2 . In the proposed DNN model, a basic building block is composed of a fully connected (FC) layer, a batch normalization (BN) layer, and a rectified linear unit (ReLU) layer. The structures of the DNN models and the training process will be discussed in the following subsections.

4.1. Basic DNN Module

In the DNN model, a basic building block is composed of a fully connected (FC) layer, a batch normalization (BN) layer, and a rectified linear unit (ReLU) layer.
In the FC layer, the input vector is multiplied with a weight matrix, and the result is added with a bias vector to form the output vector. Let x and y FC denote the N in -sized input vector and the N out -sized output vector of the FC layer, respectively. Then, the relationship between x and y FC is given by y FC = W x + b , where W is the N out × N in -sized weight matrix and b is the N out -sized bias vector.
Then, y FC passes the BN layer, which normalizes the data by re-centering and re-scaling in a mini-batch manner. To be specific, during the training process, multiple samples are fed into the DNN simultaneously as a mini-batch to improve the training efficiency and accelerate the converging. The function of the BN layer is normalizing the input data by its mean and variance. Let y BN denote the N out -sized output vector of the BN layer. Then, y BN = γ y out μ B σ B 2 + ϵ 0 + β , where μ B and σ B 2 are the mean and variance of the input mini-batch, ϵ 0 is an arbitrarily small constant for numerical stability, and γ and β are learnable parameters of the BN layer. By normalizing the distribution of the input data, the BN layer can help solve the vanishing gradient problem and avoid overfitting.
Finally, y BN enters the ReLU layer, which performs nonlinear transformation to enhance the learning capacity of DNN, where y RelU = max y BN , 0 .

4.2. DNN Model for Resource Allocation

In this subsection, we propose a DNN based method to solve the sum rate maximization problem P 1 . The network structure is shown in Figure 2. The N × N × K -sized channel matrix is flattened into a channel vector of length N 2 K in the reshaping stage. In the normalization stage, the channel vector is first converted to the dB scale and then normalized to have zero mean and unit variance.
After reshaping and normalization, the channel information enters the feature network, which is composed of M F DNN blocks, each of which consists of Q F neurons. The function of the feature network is encoding the channel information into deep features, which can be leveraged by the channel network and power network.
The channel network is composed of M C DNN blocks and a softmax layer, which takes the output of the feature network as the input. Each DNN block consists of Q C neurons. The output vector of the last DNN block is N K -sized, which is divided into NK-sized vectors and fed into the partial softmax layer. Let y n k denote the k-th element in the n-th group. After passing the softmax layer, the probability value corresponding to y n k is given by
y ^ n k = e y n k k = 1 K e y n k ,
which denotes the probability that D2D pair n chooses channel k for data transmission. In the evaluating phase, each D2D pair will choose the channel with the largest probability value, the index of which is converted into a one-hot vector representing the channel selection result.
Composed of M P DNN blocks and a softmax layer, the power network has a structure similar to the channel network. Each DNN block consists of Q P neurons. The output vector of the last DNN block is N L -sized, which is divided into NL-sized vectors and fed back into the partial softmax layer. Let p n l denote the l-th element in the n-th group. After passing the softmax layer, the probability value corresponding to p n l is given by
p ^ n l = e p n l l = 1 L e p n l ,
which denotes the probability that D2D pair n chooses power level l for data transmission. In the evaluating phase, each D2D pair will choose the power level with the largest probability.
The whole network is trained in an unsupervised manner, where the inverse sum rate of all D2D pairs is taken as the loss function
L 1 = n = 1 N R n .
During the training phase, the channel selection and power allocation based on the network output is different from that of the evaluating phase. In fact, the discretized channel selection and power allocation result cannot be applied to the loss function, because the gradient cannot be calculated during the backpropagation. To address this problem, instead of choosing the channel with the largest y ^ n k , D2D pair n directly chooses the channel as ρ n k = y ^ n k . Here, ρ n k represents the proportion of power that D2D pair n allocates to the k-th channel. On the other hand, the transmit power of D2D pair n is given by p n = l = 1 L p ^ n k P l , where p ^ n l is regarded as the weight of the l-th power level. Finally, we discuss the computational complexity of the DNN model by calculating the number of multiplications during the forward propagation. In fact, the number of multiplications of a basic DNN module is N out × N in . Therefore, the overall number of multiplications is M F Q F 2 + M C Q C 2 + M P Q P 2 .

4.3. DNN Model for Resource Allocation with Minimum Rate Constraint

In this subsection, we propose a DNN-based method to solve the sum rate maximization problem P 2 , which has a minimum rate constraint. The network structure is shown in Figure 3. An integrated network is composed of a feature network, a channel network, and a power network. In network 1, the pre-processed channel information is taken as the input, while the channel information together with the output of the previous network is taken as the input in the following networks. The structure of network 2-M is the same as that of network 1, which is omitted in the figure for simplicity. The only difference is that the input data of network 2-M is composed of the channel gain vector, the output of the previous channel network, and the output of the previous power network.
The network is trained in an unsupervised manner, where the inverse sum rate of all D2D pairs with minimum rate penalty is taken as the loss function
L 2 = n = 1 N R n η max ( R n m i n R n , 0 ) ,
where η > 0 is a weight constant to realize the tradeoff between the sum rate and the minimum rate constraint. Similar to the SAP-based method, the minimum rate constraint is better guaranteed as η increases, while the sum rate decreases because the interference becomes more severe. The channel selection and power allocation based on the network output are the same as the DNN model for problem P 1 in the previous subsection. During the training phase, the whole network is trained step by step. To be specific, network 1 is first trained by calculating loss L 2 using the output of network 1. After that, network 1 and network 2 are trained together by calculating loss L 2 using the output of network 2. The following networks are trained in this manner as well. In this way, the loop-like structure of the network can better deal with the minimum rate constraint of the D2D pairs. To satisfy the minimum rate constraint more strictly, the channel selection and power allocation result of the network is further adjusted by the SAP-based method with a small number of iterations. The computational complexity is approximately M times the complexity of the DNN model for problem P 1 . To be specific, the overall number of multiplications is approximately M ( M F Q F 2 + M C Q C 2 + M P Q P 2 ) .

5. Simulation Results

In this section, simulations are conducted to validate the performance of the proposed DNN-based resource allocation algorithm for D2D networks. The D2D pairs are randomly distributed in an area of 100 m × 100 m. The maximum distance between the transmitter and the receiver of each D2D pair is set to 15 m. The path-loss model is adopted as 128.1 + 37.6 log d [ km ] , the standard deviation of the log-normal shadowing distribution is 8 dB, and the Rayleigh fading is also adopted. Before training the network, 10,000 channel samples are generated, which are divided in a train set and a test set with the ratio of 4:1. The learning rate during training is set to be 0.001 and the Adam optimizer is adopted. Other simulation parameters are listed in Table 1. The URLLC-related parameters are the slot length T, the reliability constraint ϵ , and the minimum rate R min . These parameters are set according to [3,4], where small packets in the order of 100 bits must be delivered within no more than 1 ms and the packet error probability is no larger than 10 5 . Other parameters follow the simulation settings of the previous works [23,24].
The convergence curve of the proposed DNN-based methods is shown in Figure 4, where the number of channels is set to 5. The performance of the SAP-based method is tested as the baseline. It can be observed that the sum rates of the DNN-based methods manage to converge after training for around 10 epochs. For the sum rate maximization problem P 1 , the highest sum rate of the DNN-based method is slightly lower than that of the SAP-based method. For problem P 2 with a minimum data rate constraint, the sum rates of both methods are much lower than those of problem P 1 . This is because the minimum data rate constraint leads to more active D2D pairs, and the sum rate performance degrades due to higher interference. It is noted that the sum rate of the DNN-based method for problem P 2 first increases steadily, then decreases gradually, and finally approaches that of the SAP-based method. The reason is that the minimum rate constraint is more difficult to learn than the sum rate performance. Therefore, the DNN model first learns to improve the sum rate performance and then learns to satisfy the minimum rate constraint gradually. On the other hand, the sum rate of the SAP-based method is lower than that of the DNN-based method, because the minimum rate constraint is more strictly satisfied in the SAP-based method due to its iterative structure, which leads to higher interference among D2D pairs.
Figure 5 demonstrates the sum rate performance of the proposed methods for problem P 1 with different numbers of channels. As the channel number K increases, the sum rates of both methods increase. We can also observe that the SAP-based method provides a higher sum rate than the DNN-based method. Figure 6 depicts the complementary cumulative distribution functions (CCDFs) of the transmission rate for problem P 2 . Here, the minimum rate R min is set to 0.1 Mbps, and the channel number K is 5. From the figure, about 80% of D2D pairs can achieve a transmission rate larger than R min in the SAP-based method. In comparison, the proportion in the DNN-based method is about 75%. As the transmission rate increases, the gap between two methods becomes smaller. From Figure 5 and Figure 6, the SAP-based method achieves better performance than the DNN-based method. However, the gap between the performance of two methods is very small, and the DNN-based method is more time-efficient due to the iteration-free structure.
Table 2 compares the SAP-based method and the DNN-based method with regard to the sum rate performance and the rate excess probability. To be specific, the rate excess probability is the probability that the rate of a D2D pair is larger than R min . From the table, it is observed that the sum rate of the SAP-based method is higher than that of the DNN-based method for Problem P 1 . In comparison, the sum rate of the SAP-based method is lower than that of the DNN-based method for Problem P 2 . This is because the SAP-based method can better guarantee the minimum rate requirement due to its iterative structure, which results in the rate loss compared with the DNN-based method.

6. Conclusions

This paper develops a novel resource allocation algorithm for overlay D2D networks based on DNN to support URLLC services. We first formulate two sum rate maximization problems with and without minimum rate constraint. Then, we design an iterative algorithm based on game theory as the baseline. After that, a DNN-based channel selection and power allocation method is developed. The network is trained in an unsupervised manner with no need for labeled data, and two different loss functions are designed for the two formulated problems, respectively. Finally, the performance of the proposed methods is demonstrated through numerical simulation.
In fact, the proposed method could be adapted from different aspects so that it could be applied in a practical system. For example, transfer learning could be applied to accelerate the training process under different scenarios. The detailed implementation of the proposed methods in a real D2D 5G environment will be investigated in our future work.

Author Contributions

Conceptualization, Z.Z. and Y.C.; methodology, Z.Z. and Y.C.; software, G.D.; validation, G.D.; formal analysis, G.D.; investigation, G.D.; resources, G.D.; data curation, G.D.; writing—original draft preparation, G.D.; writing—review and editing, G.Y.; visualization, G.D.; supervision, Z.Z. and Y.C.; project administration, Z.Z. and Y.C.; funding acquisition, Z.Z. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by The Laboratory Open Fund of Beijing Smart-chip Microelectronics Technology Co., Ltd., Beijing, China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, S.; Xu, X.; Wu, Y.; Lu, L. 5G: Towards energy-efficient, low-latency and high-reliable communications networks. In Proceedings of the 2014 IEEE International Conference on Communication Systems, Macau, China, 19–21 November 2014; pp. 197–201. [Google Scholar]
  2. Chen, H.; Abbas, R.; Cheng, P.; Shirvanimoghaddam, M.; Hardjawana, W.; Bao, W.; Li, Y.; Vucetic, B. Ultra-reliable low latency cellular networks: Use cases, challenges and approaches. IEEE Commun. Mag. 2018, 56, 119–125. [Google Scholar] [CrossRef] [Green Version]
  3. Parvez, I.; Rahmati, A.; Guvenc, I.; Sarwat, A.I.; Dai, H. A survey on low latency towards 5G: RAN, core network and caching solutions. IEEE Commun. Surveys Tuts. 2018, 20, 3098–3130. [Google Scholar] [CrossRef]
  4. Service Requirements for Cyber-Physical Control Applications in Vertical Domains, Document TS 22.104; Version 18.2.0; 3rd Generation Partnership Project (3GPP); 2021.
  5. Soret, B.; Mogensen, P.; Pedersen, K.I.; Aguayo-Torres, M.C. Fundamental tradeoffs among reliability, latency and throughput in cellular networks. In Proceedings of the 2014 IEEE Globecom Workshops (GC Wkshps), Austin, TX, USA, 8–12 December 2014; pp. 1391–1396. [Google Scholar]
  6. Tehrani, M.N.; Uysal, M.; Yanikomeroglu, H. Device-to-device communication in 5G cellular networks: Challenges, solutions, and future directions. IEEE Commun. Mag. 2014, 52, 86–92. [Google Scholar] [CrossRef]
  7. Asadi, A.; Wang, Q.; Mancuso, V. A survey on device-to-device communication in cellular networks. IEEE Commun. Surveys Tuts. 2014, 16, 1801–1819. [Google Scholar] [CrossRef] [Green Version]
  8. Yu, G.; Xu, L.; Feng, D.; Yin, R.; Li, G.Y.; Jiang, Y. Joint mode selection and resource allocation for device-to-device communications. IEEE Trans. Commun. 2014, 62, 3814–3824. [Google Scholar] [CrossRef]
  9. Jiang, Y.; Liu, Q.; Zheng, F.; Gao, X.; You, X. Energy-efficient joint resource allocation and power control for D2D communications. IEEE Trans. Veh. Technol. 2016, 65, 6119–6127. [Google Scholar] [CrossRef] [Green Version]
  10. Dai, J.; Liu, J.; Shi, Y.; Zhang, S.; Ma, J. Analytical modeling of resource allocation in D2D overlaying multihop multichannel uplink cellular networks. IEEE Trans. Veh. Technol. 2017, 66, 6633–6644. [Google Scholar] [CrossRef]
  11. Ma, R.; Xia, N.; Chen, H.; Chiu, C.; Yang, C. Mode selection, radio resource allocation, and power coordination in D2D communications. IEEE Wireless Commun. 2017, 24, 112–121. [Google Scholar] [CrossRef]
  12. Zhou, Z.; Ota, K.; Dong, M.; Xu, C. Energy-efficient matching for resource allocation in D2D enabled cellular networks. IEEE Trans. Veh. Technol. 2017, 66, 5256–5268. [Google Scholar] [CrossRef] [Green Version]
  13. Lyu, J.; Chew, Y.H.; Wong, W.C. A stackelberg game model for overlay D2D transmission with heterogeneous rate requirements. IEEE Trans. Veh. Technol. 2016, 65, 8461–8475. [Google Scholar] [CrossRef] [Green Version]
  14. Chu, Z.; Yu, W.; Xiao, P.; Zhou, F.; Al-Dhahir, N.; Ul, Quddus, A.; Tafazolli, R. Opportunistic spectrum sharing for D2D-based URLLC. IEEE Trans. Veh. Technol. 2019, 68, 8995–9006. [Google Scholar] [CrossRef]
  15. Wu, Y.; Wu, D.; Ao, L.; Yang, L.; Fu, Q. Contention-based radio resource management for URLLC-oriented D2D communications. IEEE Trans. Veh. Technol. 2020, 69, 9960–9971. [Google Scholar] [CrossRef]
  16. Wu, Y.; Wu, D.; Yue, C.; Yang, Y. Joint Reservation and Contention-Based Access for URLLC-Enabled D2D Communications. IEEE Commun. Lett. 2021, 26, 212–216. [Google Scholar] [CrossRef]
  17. Liu, L.; Yu, W. A D2D-Based Protocol for Ultra-Reliable Wireless Communications for Industrial Automation. IEEE Trans. Wireless Commun. 2018, 17, 5045–5058. [Google Scholar] [CrossRef]
  18. Yuan, Y.; Yang, T.; Hu, Y.; Feng, H.; Hu, B. Two-Timescale Resource Allocation for Cooperative D2D Communication: A Matching Game Approach. IEEE Trans. Veh. Technol. 2021, 70, 543–557. [Google Scholar] [CrossRef]
  19. Lee, M.; Xiong, Y.; Yu, G.; Li, G.Y. Deep neural networks for linear sum assignment problems. IEEE Wireless Commun. Lett. 2018, 7, 962–965. [Google Scholar] [CrossRef]
  20. Zappone, A.; Sanguinetti, L.; Debbah, M. User association and load balancing for massive MIMO through deep learning. In Proceedings of the 2018 52nd Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 28–31 October 2018; pp. 1262–1266. [Google Scholar]
  21. Gao, J.; Hu, M.; Zhong, C.; Li, G.; Zhang, Z. An attention-aided deep learning framework for massive MIMO channel estimation. IEEE Trans. Wireless Commun. 2021. early access. [Google Scholar] [CrossRef]
  22. Qin, Z.; Ye, H.; Li, G.Y.; Juang, B.F. Deep learning in physical layer communications. IEEE Wireless Commun. 2019, 26, 93–99. [Google Scholar] [CrossRef] [Green Version]
  23. Lee, W. Resource allocation for multi-channel underlay cognitive radio network based on deep neural network. IEEE Commun. Lett. 2018, 22, 1942–1945. [Google Scholar] [CrossRef]
  24. Tan, J.; Liang, Y.C.; Zhang, L.; Feng, G. Deep reinforcement learning for joint channel selection and power control in D2D networks. IEEE Trans. Wireless Commun. 2021, 20, 1363–1378. [Google Scholar] [CrossRef]
  25. Polyanskiy, Y.; Poor, H.V.; Verdu, S. Channel coding rate in the finite blocklength regime. IEEE Trans. Inf. Theory 2010, 56, 2307–2359. [Google Scholar] [CrossRef]
  26. Schiessl, S.; Gross, J.; Al-Zubaidy, H. Delay analysis for wireless fading channels with finite blocklength channel coding. In Proceedings of the 18th ACM International Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems, New York, NY, USA, 2 November 2015; pp. 13–22. [Google Scholar]
  27. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
Figure 1. The system model of D2D networks.
Figure 1. The system model of D2D networks.
Sensors 22 01551 g001
Figure 2. The DNN model for problem P 1 .
Figure 2. The DNN model for problem P 1 .
Sensors 22 01551 g002
Figure 3. The DNN model for problem P 2 .
Figure 3. The DNN model for problem P 2 .
Sensors 22 01551 g003
Figure 4. Convergence curve of the DNN-based methods.
Figure 4. Convergence curve of the DNN-based methods.
Sensors 22 01551 g004
Figure 5. Sum rate performance for problem P 1 with varying channel number.
Figure 5. Sum rate performance for problem P 1 with varying channel number.
Sensors 22 01551 g005
Figure 6. CCDF of the transmission rate for problem P 2 .
Figure 6. CCDF of the transmission rate for problem P 2 .
Sensors 22 01551 g006
Table 1. Simulation parameters.
Table 1. Simulation parameters.
ParameterValue
Number of D2D pairs, N10
Number of channels, K2,4,5,6,7
Channel bandwidth, B180 KHz
Slot length, T1 ms
Maximum transmit power, P max 23 dBm
Power spectral density of noise, N 0 −174 dBm/Hz
Reliability constraint, ϵ 10 5
Minimum rate, R min 10 5 bps
Initial exploration probability μ 0.5
Decay rate w0.9
Table 2. Performance comparison of the SAP-based method and the DNN-based method.
Table 2. Performance comparison of the SAP-based method and the DNN-based method.
AlgorithmSum Rate of P 1Sum Rate of P 2Rate Excess Probability of P 2
DNN-based Method23.84 Mbps21.90 Mbps0.7382
SAP-based Method24.32 Mbps21.52 Mbps0.7903
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, Z.; Chi, Y.; Ding, G.; Yu, G. Deep-Learning-Based Resource Allocation for Time-Sensitive Device-to-Device Networks. Sensors 2022, 22, 1551. https://doi.org/10.3390/s22041551

AMA Style

Zheng Z, Chi Y, Ding G, Yu G. Deep-Learning-Based Resource Allocation for Time-Sensitive Device-to-Device Networks. Sensors. 2022; 22(4):1551. https://doi.org/10.3390/s22041551

Chicago/Turabian Style

Zheng, Zhe, Yingying Chi, Guangyao Ding, and Guanding Yu. 2022. "Deep-Learning-Based Resource Allocation for Time-Sensitive Device-to-Device Networks" Sensors 22, no. 4: 1551. https://doi.org/10.3390/s22041551

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop