Next Article in Journal
Effects of Charge Traps on Hysteresis in Organic Field-Effect Transistors and Their Charge Trap Cause Analysis through Causal Inference Techniques
Next Article in Special Issue
A Cost-Effective Lightning Current Measuring Instrument with Wide Current Range Detection Using Dual Signal Conditioning Circuits
Previous Article in Journal
Extracting Traffic Signage by Combining Point Clouds and Images
Previous Article in Special Issue
Ambient Hot Box: An Instrument for Thermal Characterization of Building Elements and Constructive Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Interference Discrimination Testbed Employing AI Ensemble Algorithms for 6G TDD Networks

1
Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China
2
School of Information Science and Technology, ShanghaiTech University, Shanghai 201210, China
3
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
4
School of Microelectronics, Shanghai University, Shanghai 200444, China
5
Shanghai Frontier Innovation Research Institute, Shanghai 201100, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2264; https://doi.org/10.3390/s23042264
Submission received: 14 January 2023 / Revised: 6 February 2023 / Accepted: 15 February 2023 / Published: 17 February 2023

Abstract

:
The Internet-of-Things (IoT) massive access is a significant scenario for sixth-generation (6G) communications. However, low-power IoT devices easily suffer from remote interference caused by the atmospheric duct under the 6G time-division duplex (TDD) mode. It causes distant downlink wireless signals to propagate beyond the designed protection distance and interfere with local uplink signals, leading to a large outage probability. In this paper, a remote interference discrimination testbed is originally proposed to detect interference, which supports the comparison of different types of algorithms on the testbed. Specifically, 5,520,000 TDD network-side data collected by real sensors are used to validate the interference discrimination capabilities of nine promising AI algorithms. Moreover, a consistent comparison of the testbed shows that the ensemble algorithm achieves an average accuracy of 12% higher than the single model algorithm.

1. Introduction

Massive access is defined as a typical scenario of sixth-generation (6G) communications by IMT-2030 Promotion Group. Numerous Internet of Things (IoT) devices will be connected to the communication network [1]. However, the remote interference caused by the atmospheric duct brings about the interference signal exceeding the guard period (GP), which interferes with the co-frequency uplink signal reception of low-power IoT devices in 6G time-division duplex (TDD) networks and increases the risk of communication interruption for mobile users.
The TDD mode, which prominently suffers from the interference of the atmospheric duct, refers to the uplink and downlink utilizing the same frequency band to transmit information at different times [2]. The GP, as shown in Figure 1, is applied to protect the uplink signal from the interference of the downlink signal [3]. The interference signal can be filtered by the sensor within the GP protection range. However, the distance of remote interference will far exceed this range. The atmospheric duct, which results from non-standard meteorological conditions, captures the electromagnetic wave and induces the signal to propagate in the ducting layer [4]. The atmospheric duct captures the signal and allows the signal to propagate beyond the GP maximum protection distance with low path loss [5]. Thus, the captured signal maintains a high signal strength and interferes with the uplink signal reception of remote IoT devices [6].
According to statistics, China, Japan, Netherlands, and the United States have suffered from the interference of the atmospheric duct for a long time [7,8,9,10]. In the process of 5G research, remote interference has attracted the attention of researchers. 3GPP promoted a remote interference project in the standardization research of 5G-Beyond to analyze the adverse impact of the remote interference on communication systems [11]. Moreover, ITU recommended that the atmospheric duct should be considered in channel modeling [12]. Ericsson and other telecom companies began monitoring and analyzing the effects of remote interference on communication systems [13]. Clearly, the negative effects of the atmospheric duct have attracted the attention of many researchers. In the future research of 6G communications, remote interference will also be an unavoidable problem.
Currently, there are two main methods used to detect and estimate the atmospheric duct: (1) using meteorological factors to calculate the atmospheric refractive index [14]; (2) using radar to measure the atmospheric duct [15]. Nonetheless, Method (1) usually utilizes PETOOL platform to generate simulation data [16]. It contains several assumptions to generate simulation data, which hardly reflect reality. Method (2) is suitable for ocean scenarios and is expensive. In addition, remote interference typically occurs on land where water vapor evaporation is large. Considering complex scenarios and large amounts of data, traditional modeling methods do not work.
Motivated by the above challenges, a remote interference discrimination testbed employing AI ensemble algorithms for 6G wireless communications is proposed. The contributions of this paper are summarized as follows:
  • A remote interference discrimination testbed is originally proposed, which adopts 5,520,000 TDD network-side interfered data to discriminate the remote interference. A large number of measurement data could effectively appraise the interference discrimination ability of different AI algorithms;
  • The testbed verifies the interference discrimination ability of two types of a total of nine AI algorithms, which lays the foundation for the application of the testbed in different hardware environments;
  • According to the consistent comparison, numerical results illustrate that the ensemble algorithm achieves an average accuracy of 12% higher than the single model algorithm. The work fills the gap of remote interference in the 6G communication scenario and helps mobile operators improve network optimization capabilities under remote interference.
The remainder of the paper is organized as follows. In the next section, the recent studies of atmospheric duct and the framework of the proposed testbed are introduced. Section 3 shows the employed ensemble discriminant algorithms. Extensive experiments are presented in Section 4. Finally, the conclusions are summarized in Section 5.

2. Related Work and Testbed Design

2.1. Related Work

While most of the existing research literature on the atmospheric duct has focused on calculating the height of the ducting layer, there has been little analysis of the interference discrimination in communication systems. Currently, there are two main approaches to detect and estimate the atmospheric duct, including theoretical calculations and practical measurements.
Ray-optics (RO) method and parabolic equation (PE) method are developed to calculate the trajectory of the ducting layer. For example, a RO method was applied to calculate ray trajectories with atmospheric ducts in Ref. [17]. The authors analyzed delay spreads to determine the fading behavior of the channel, which compensated for a realistic analysis for the delay spread of ducting channels. A PE-based tool (PETOOL) was developed in Ref. [18], who analyzed the ideal ducting effect from 800 MHz to 20 GHz.
Considering the interference of the duct on the electromagnetic wave signal, some studies utilized radar and other equipment for measurement. In Ref. [19], a comprehensive observation experiment was carried out in the Guangdong Province of China. A shore-based navigation radar was used for over-the-horizon detection and radiosondes were used to measure the atmospheric profile. A method of detecting atmospheric ducts using a wind profiler radar and a radio acoustic sounding system was proposed in Ref. [20]. The measurements were carried out in the Liaoning Province of China. These activities all take place at sea, and the expensive cost and restrictions hinder land measurement.

2.2. Testbed Design

The proposed remote interference discrimination testbed is shown in Figure 2. It consists of four modules, including meteorology and signal module, data processing module, AI-based learning module, and validation module.
First of all, the meteorology and signal module adopts sensors to collect meteorological and network-side data. Secondly, in the data processing module, the collected data is cleaned and divided into two parts: meteorological factors and network factors. Then, the factors are input into AI-based learning module to acquire data characteristics. Finally, the validation module uses the measurement data to verify the interference discrimination ability of the model. Our previous work has completed the meteorology and signal module, and validation module [21]. In the following, we focus on introducing the data processing module and AI-based learning module.
Without loss of generality, a channel with atmospheric duct interference is considered. In data processing, interference discrimination requires elucidating which factors are relevant for the wireless channel under ducting interference. The contributory factors are deduced in the following, which consists of meteorological factors and network factors.

2.2.1. Meteorological Factors

Atmospheric refraction is the bending of electromagnetic waves propagating in the atmospheric media. The degree of refraction could be described by the refractive index, which is expressed as [17]
n = c v ,
where c represents the light speed, and v refers to the velocity of the electromagnetic wave in the medium. The atmospheric refractivity is employed to replace the refractive index due to the minuscule value of n being ignored when calculated for most cases [22]. The refractivity can be described as [12]
N = ( n 1 ) × 10 6 = 77.6 T × ( p + 4810 e T ) ,
where T denotes the temperature, p represents the atmospheric pressure, and e indicates the vapor pressure.
Notably, the curvature of the earth needs to be considered since the signal captured by the atmospheric duct is capable of traveling long distances. As a result, the modified refractivity, which considers the curvature of the earth, can be expressed as [12]
M = N + h r e × 10 6 ,
where h denotes the height above ground, and r e is the earth radius. The atmospheric duct occurs when d M d h < 0 . The appearance of the atmospheric duct is related to meteorological parameters, whose changes are inseparable from time.

2.2.2. Network Factors

The PE method, utilizing paraxial approximation of the Helmholtz equation, could model the changes of refractivity in the atmosphere and simulate complex boundary conditions. As such, the PE-based path loss model, which integrates diverse conditions well, can be represented as [23]
L p ( z , h ) = 20 lg | u ( z , h ) | + 20 lg ( 4 π ) + 10 lg ( z ) 30 lg ( λ ) ,
where L p denotes the path loss of the signal, z represents the horizontal distance of signal propagation, λ is the carrier wavelength, and u refers to the field strength, which can be written as [23]
u = 2 π + B ( θ ) e 2 i n p b h d p b ,
s i n ( θ ) = λ p b ,
where B refers to the beam function, θ denotes the down tilt angle, and p b indicates the beam. When the antenna is modeled as a Gaussian function, B can be formulated as [23]
B ( θ ) = A e ( 2 lg 2 θ 2 β 2 ) ,
where A denotes the normalization constant, and β refers to the half-power beamwidth.
Under these circumstances, the initial field strength can be written as [23]
u ( 0 , h ) = A k β 2 2 π lg 2 e i k θ h e β 2 8 lg 2 k 2 ( h h a ) 2 ,
where k indicates the incident wave beam, and h a represents the antenna height.
The solution of the field strength can be described as [23]
u h ( z , h ) = i k ( 1 k 2 2 h 2 + n 2 ( 1 + h r e ) 2 1 ) u ( z , h ) .
Equation (9) needs to be solved by the Fourier transform and inverse transform. The relationship between field strengths can be expressed as [23]
U ( z , p b ) = + u ( z , h ) e 2 i π p b h d h ,
u ( z , h ) = + U ( z , p b ) e 2 i π p b h d p b .
After finishing the Fourier transform, the increment can be calculated as [23]
u ( z + Δ z , h ) = e i k Δ z ( 1 k 2 2 h 2 + n 2 ( 1 + h r e ) 2 1 ) u ( z , h ) .
As can be seen from the above analysis, the PE method adopts the split-step Fourier transform to solve the equation due to the complex nonlinear relationship between the path loss of the signal and contributory factors. In summary, the contributory factors of the atmospheric duct include temperature, atmospheric pressure, relative humidity, time, longitude, latitude, antenna height, and down tilt angle. These factors mentioned above affect the path loss of the signal.
Considering the contributory factors, the corresponding data is selected from the dataset. Traditional modeling methods struggle to effectively learn and represent data features in the presence of huge amounts of data, so AI-based learning methods have emerged as a promising solution.

3. AI-Based Discriminant Algorithms

The processed data is input to the AI-based learning module to generate the feature model. The model can be adopted to discriminate the remote interference and warn the operator to operate to avoid remote interference. Obviously, an accurate model is crucial for the interference discrimination framework. The discriminant algorithm is mainly separated into two parts, including the single model algorithms, and the ensemble algorithms [24]. The details of the discriminant algorithms are as follows.

3.1. Single Model Algorithms

The single model algorithms have been applied in many fields. Some investigations have verified that some single model algorithms have pleasant performance in remote interference discrimination, which is the focus of the subsection.
Most single model algorithms adopt mathematical expressions to judge categories. For example, nearest distance matching, distribution model matching, and so on. Single model algorithms often achieve satisfactory performance in communication problems such as low interference channel estimation [25]. The channel contributory factors of interference discrimination exist as complex nonlinear relationships, and require a high demand for single model algorithms. The single model algorithms, which have been employed for interference discrimination, will be introduced as follows [26].

3.1.1. kNN

The k-Nearest Neighbors (kNN) algorithm is an earlier supervised machine learning algorithm. The keystone of kNN is using k adjacent values to represent sample points [27]. The category of sample points is determined by the k nearest neighbors, which is the same as the majority of the neighbors. Many ways can be applied to express the distance between points, including the Euclidean distance, Manhattan distance, cosine distance, Chebyshev distance, and so forth [28].
The Euclidean distance is often selected as the calculation index, which can be expressed as [28]
d euc = i = 1 m ( x i y i ) 2 ,
where m indicates the data dimension. With the increase of variables, the distinguishing ability of Euclidean distance becomes worse.
The Manhattan distance is written as [28]
d man = i = 1 m ( | x 1 i y 1 i | + | x 2 i y 2 i | ) .
The Manhattan distance has a fast calculation speed, but when the differences of variables are large, some features will be ignored.
The cosine distance is represented as [28]
d cos = i = 1 m ( x 1 i y 1 i + x 2 i y 2 i x 1 i 2 + x 2 i 2 × y 1 i 2 + y 2 i 2 ) .
The cosine distance is suitable for many variables and solving the problems of outliers and sparse data, whereas it discards the useful information contained in the vector length.
The Chebyshev distance is executed as [28]
d che = max ( | x i y i | ) .
The Chebyshev distance is generally utilized to calculate the sum of distances, such as the logistic store.

3.1.2. SVM

The support vector machine (SVM) is a supervised learning algorithm, which especially supports the binary classification. SVM maps samples into space and finds a hyperplane to maximize the interval between samples. The classification of training samples is divided into two parts, including linear and nonlinear. The linear data could be divided into positive and negative samples [29]. SVM uses a hyperplane to divide the positive and negative samples. The selection of the hyperplane is shown in Figure 3, which can be described as [30]
ω x i + b = 0 ,
where ω denotes the normal vector, and b indicates the distance between the plane and coordinate origin.
Building an optimized hyperplane in a complex nonlinearly separable problem is done using kernels. The kernel functions are of many types such as Gaussian, polynomial, sigmoid, Cauchy, and so on [31]. Kernel functions map linearly inseparable data to high-dimensional space.
The Gaussian kernel function is performed as [32]
k gau ( x , y ) = e | | x y | | 2 2 σ 2 ,
where σ represents the standard deviation. The Gaussian kernel function is commonly used in SVM, and the essence of Gaussian is to map each sample point to an infinite-dimensional feature space, which means the deformation of samples is extremely complex, but the characteristics of each sample are clear.
The polynomial function is denoted by [32]
k pol ( x , y ) = ( x · y + 1 ) D ,
where D denotes the degree of the polynomial. The function indicates the similarity of vectors in the training set. The polynomial function is relatively stable, but it involves many parameters.
The sigmoid function is defined as [32]
k sig ( x , y ) = t a n h ( x · y + 1 ) .
Sigmoid is an S-shaped function, which is often employed as the activation function of the neural network to map variables between 0 and 1.
The Cauchy function is written as [32]
k cau ( x , y ) = 1 | | x y | | 2 σ + 1 .
The Cauchy function is mainly applied to deal with high-dimensional data.

3.1.3. NB

Naive Bayes (NB) is a discriminant method based on Bayesian theorem and feature condition independence hypothesis [33]. The advantage of NB is that it combines the prior probability and the posterior probability, that is, it avoids the subjective bias of using only the prior probability and the over fitting phenomenon of using sample information alone [34]. However, NB requires few estimated parameters, it is not sensitive to missing data, and the assumption is relatively simple, so the accuracy of the algorithm is affected. According to different assumptions, NB includes Gaussian NB (GNB), Multinomial NB (MNB), Complement NB (CNB), Bernoulli NB (BNB), Categorical NB, and so on [35].
GNB denotes the prior distribution, which is assumed to be Gaussian [36]. BNB is designed for binary discrete data [37]. The Categorical NB assumes that each feature described by the index has its own classification distribution [38]. MNB is utilized to calculate the probability of discrete features [39]. CNB can be used to classify imbalanced datasets when the features do not satisfy the conditions of mutual independence. [40]. NB contains multiple input variables and target variables as model outputs. Let S be the state of the variable and X = ( x 1 , x 2 , . . . , x n ) be the state of n input features. To estimate the value of S based on X, the conditional probability of S needs to be calculated by X, and the expression is [41]
p ( S | X ) = p ( X | S ) p ( S ) p ( X ) ,
where p ( S ) and p ( X ) are constants that are obtained from data. p ( X | S ) can be calculated as [41]
p ( X | S ) = p ( x 1 , x 2 , . . . , x n | S ) = i = 1 n p ( x i | S ) .
The expression of p ( S | X ) can be simplified as [41]
p ( S | X ) = p ( S ) p ( X ) i = 1 n p ( x i | S ) .

3.2. Ensemble Algorithms

As one of the current research hotspots, ensemble learning has been applied tentatively in many fields, such as image processing, malware detection, and so on [42]. The multi-model properties of ensemble learning enable to avoid the imprecise characteristic of a single model, which also shows potential in solving complex problems.
Ensemble learning refers to strategically generating multiple weak classifiers and then combining them into a strong classifier to complete the discrimination task, which has superior generalization ability. Next, several effective algorithms in some fields will be introduced. The ensemble algorithms are mainly divided into two categories, including serial and parallel algorithms [43]. Random Forest (RF) and Bootstrap Aggregating (Bagging) belong to the parallel algorithms. Boosting and Stacked Generalization (Stacking) are parts of the serial algorithms.

3.2.1. RF

RF is a classifier containing multiple decision trees, and its output category is determined by the mode of the category output by individual decision trees [44]. The decision tree adopts the top-down recursive method, which constructs a tree with the fastest entropy decline based on information entropy. The information entropy is defined as [45]
H = i = 1 n p i ln ( p i ) ,
where H refers to the information entropy, and p indicates the probability. It can be seen from Figure 4 that RF consists of multiple decision trees. Each decision tree will get a discrimination result, and all the results determine the final output. The advantage of RF is that it is able to process high-dimensional data and find the relationship between different variables [46]. The advantage of RF is that it can process high-dimensional data, has strong anti-noise ability, and avoids the overfitting problem.
RF has superior performance in numerous aspects, especially in pathological research and financial investment. However, because of its slow pace, the random forest classifier is not applicable to real-time predictions.

3.2.2. Bagging

Bagging is an algorithm framework, which trains several different models respectively, and then lets all models vote to test the output of samples [47]. As shown in Figure 5, Bagging adopts a sampling with replacement to generate multiple training subsets, which are employed to train classifiers [48]. Each training process is independent, so the process could be accelerated by parallel computing [49]. Especially, the training subset of Bagging is randomly selected, which means that different subsets may contain the same data. Moreover, Bagging introduces randomization in the training of each classifier. After training, all classifiers are combined to reduce the variance of prediction results. After the L-th iteration, the expectation of the strong classifier is expressed as [50]
ϕ ( x ) = E L ( x , L ) .
The difference between the real value y and the predicted value of the weak classifier can be written as [50]
E L ( y ϕ ( x , L ) ) 2 = y 2 2 y E L ϕ ( x , L ) + E L ϕ 2 ( x , L ) .
The comparison result of classifiers is described as [50]
E L ( y ϕ ( x , L ) ) 2 ( y ϕ ( x ) ) 2 .
The expectation of multiple weak classifiers is better than that of the strong classifier, that is, Bagging is able to effectively improve the discrimination accuracy, especially when the variance between the variables is large.

3.2.3. Boosting

Similar to Bagging, Boosting also trains multiple weak classifiers to jointly decide the final output [51]. However, weak classifiers are strengthened and trained by weighting in Boosting [52]. Boosting is a framework, which obtains the subset, and utilizes the weak classification algorithm to train to generate a series of base classifiers [53]. The optimization model of Boosting is executed as [53]
F m ( x ) = F m 1 ( x ) + argmin θ i = 1 n L ( y i , F m 1 ( x i ) + θ ( x i ) ) ,
where L denotes the greedy optimization. To solve the detailed problem of subsets and classifiers, Boosting derives multifarious algorithms, including Adaptive Boosting (AdaBoost), Gradient Boosting Decision Tree (GBDT), Xtreme Gradient Boosting (XGBoost), and so on.
AdaBoost will select the key classification feature set in the training set for many times. It trains the component weak classifier step by step and selects the best weak classifier with an appropriate threshold. Finally, the best weak classifier for each iteration is selected to construct a strong classifier. However, AdaBoost combines weak classifiers to construct a strong classifier [54]. The weights of each weak classifier are not equal, and the stronger classifier will be assigned the high weight [55]. Specifically, the weighted error of the k-th weak classifier G k ( x ) is written as [54]
e k = P ( G k ( x i ) y i ) = i = 1 m w k i I ( G k ( x i ) y i ) ,
where w indicates the output weight. The weight coefficient of the k-th G k ( x ) is defined as [54]
α k = 1 2 log ( 1 e k e k ) .
It can be found that the weight coefficient decreases with the increase of weighted error.
The expression of updated weight is [54]
w k + 1 , i = w k i exp ( α k y i G k ( x i ) ) i = 1 m exp ( α k y i G k ( x i ) ) .
AdaBoost needs a quality dataset because it is hard to handle noisy data and outliers. At present, AdaBoost is being used to classify text and images rather than binary classification problems.
The core of GBDT is that each tree learns the residual of the sum of all previous tree conclusions, which is the accumulation of the real value after adding the predicted value [56]. The fitting error of GBDT, which is replaced by the negative gradient of the loss function, is reduced by multiple iterations [57]. The negative gradient expression of the i-th sample in the t-th iteration is performed as [56]
r t i = [ L ( y i , f ( x i ) ) f ( x i ) ] f ( x ) = f t 1 ( x ) ,
where r t i denotes the negative gradient, and L represents the loss function. After getting the t-th decision tree, the optimal solution of the loss function is given by [56]
c t j = arg min x i R t j L ( y i , f t 1 ( x i ) + c ) ,
where c represents the optimal solution, R indicates the region of the child node, and j denotes the number of the child node. The optimal solution could be utilized to update the weak classifier [58].
XGBoost adopts a similar theory to GBDT [59]. GBDT applies the first derivative in the loss function, but the loss function of XGBoost is approximated by the second-order Tailor expansion. Furthermore, the objective function of XGBoost imports a regularizer to avoid the over-fitting problem, which is expressed as [60]
O b j ( t ) = i = 1 n L ( y i , y ^ i ( t ) ) + i = 1 t Ω ( f i ) ,
where y ^ denotes the forecasting sample, and Ω represents the regularizer. XGBoost employs regularization to avoid overfitting, and it usually has superior performance in dealing with small and medium datasets.

3.2.4. Stacking

Stacking is an ensemble technique that combines multiple discrimination results generated by using different learning algorithms on the dataset [61]. Stacking contains two layers of classification models, as shown in Figure 6. The first layer applies various classifiers to predict the result. The result is input into the second layer as the training set. The second layer is utilized to assign higher weights to better classifiers, so the two-layer model could effectively reduce the variance [62]. Hence, Stacking will select several classifiers with good fitting for deciding the final result. However, the good performance of a single classifier does not mean that the combined effect is ideal.

4. Interference Discrimination Experiments

In this section, practical sensors-collected remote interference measurement data is employed to analyze the testbed effectiveness. The selected single model algorithms were employed to discriminate the interference of the base station [26]. The selected ensemble algorithms have excellent performance on complex problems. Furthermore, accuracy and recall are applied to assess the performances of algorithms. Accuracy refers to the probability that the models correctly judge the test data, and recall indicates the probability that the models correctly judge the data interfered by the atmospheric duct. As such, the experiments include three parts. (a) Change the size of the dataset; (b) Change the imbalance ratio (IR) of the data size; (c) Test the robustness of algorithms; (d) Time complexity.

4.1. Interference Dataset

The dataset is the measurement of the sensor under the TDD system, which is provided by China Mobile Group Jiangsu Co., Ltd. Some base stations were interfered by the atmospheric duct in Jiangsu Province of China, which interfered with the reception of the uplink signal. The data is collected from 240,000 antennas in Jiangsu, including the longitude, latitude, time, antenna height, and down tilt angle.
Figure 7 shows the number of interfered base stations, which gradually increases from 1.00 a.m. to 7.00 a.m., with the number dropping dramatically from 8.00 a.m. The trend shows that the atmospheric duct usually appears from midnight to the morning. From the explanation of meteorology, the temperature of the ground drops quickly and the lower atmosphere is prone to temperature inversion from midnight to the morning, which means that within a certain height, the temperature increases with the vertical height, which causes the atmospheric duct phenomenon.
The meteorological data is obtained from CFSv2, which is a fully coupled model representing the interaction between the earth’s atmosphere, oceans, land, and sea ice [63]. The meteorology of CFSv2 includes the temperature, relative humidity, pressure, salinity, and so on. We download the temperature, relative humidity, and pressure data, which is related to the atmospheric duct, to match with the base station according to the longitude and latitude.

4.2. Algorithm Settings

The hardware and software configurations of experiments are listed in Table 1. The algorithms in Section 3 are selected to test the performance of the interfered dataset. Unless otherwise specified, all parameters are set to the values in Table 2 by default. The empirical results show that a large proportion of algorithms converge after 100 iterations, which is chosen as the maximum number of iterations in our experiments. Particularly, the iterations of AdaBoost are 500 because the higher iterations of the algorithm will significantly improve the discrimination results.

4.3. Sensitivity of the Algorithms to the Data Size

To verify the influence of different data sizes, the size of the training set is set as 20,000, 30,000, 40,000, 50,000, and 60,000, respectively. The IR of each training set is 5:1: for instance, in the training set of 20,000, about 3333 pieces of data are interfered by the atmospheric duct, and the rest are normal. Moreover, the equivalent data is sampled per hour to form the training set.
The size of the test set is set to 20% of the total number of the training set. The number of the interfered data and the normal data are the same in the test set, which is applied to emphasize the learning ability of the algorithms for the imbalanced dataset. Similarly, the equivalent data is sampled per hour to form the test set, which ensures fairness in the time domain.
There is no overlap between the training set and the test set. When the size of the training set changes, both the training set and the test set will be selected randomly. Besides, two indicators, including accuracy and recall, are applied to evaluate the learning ability of the algorithms. The expression of accuracy can be expressed as
A c c = T P + T N T P + T N + F P + F N ,
where T P is the true positive, T N is the true negative, F P is the false positive, and F N is the false negative. In the interference discrimination problem, T P refers to the interfered samples that are judged correctly by algorithms, T N denotes the interfered samples that are judged incorrectly, F P represents the undisturbed samples that are judged correctly, and F N indicates the undisturbed samples that are judged incorrectly.
The expression of recall is defined as
R e c a l l = T P T P + F N .
The recall is utilized to reflect the judgment ability of the algorithm for specific indicators, which is especially adopted to display the judgment of the interfered data in the interference discrimination problem.
Table 3 shows the specific classification results on different datasets. The accuracy results of single model algorithms and ensemble algorithms are illustrated in Figure 8a, and the recall of two kinds of algorithms is shown in Figure 8b.
In Figure 8a, the accuracy of all algorithms keeps improving with the increase of data, which means that the amount of data has a significant impact on the accuracy. Specifically, Bagging has the highest accuracy, which demonstrates it could better characterize the complex nonlinear relationship between variables. The recall has a similar trend with the accuracy, as shown in Figure 8b, which shows that the recall of Bagging is higher than the others, that is, Bagging could well learn the characteristics of the minority in the imbalanced datasets.
Stacking, RF, and XGBoost have stationary performance on the dataset, which validates that the three algorithms could fit the complex nonlinear relationship among variables well. The accuracy of kNN is generally precise, which indicates that there are a few differences among the variables, so distance matching is hard to find the internal relationship among variables. NB only needs a few samples to achieve high accuracy, so the accuracy has changed rarely when the amount of data is sufficient. Meanwhile, the generalization ability of the model is weak, so the learning ability of the minority is poor. The accuracy of AdaBoost is not high, because the weights tend to the classifiers that have superior performance, and the generalization ability of the model is affected.
However, the accuracy results of SVM and GBDT only attain 50.00%, and the recall results of the two algorithms are almost 0.00%. It is revealed that the two algorithms judge the data as normal data with a high proportion in the training set. We also test the ideal case with a 1:1 imbalance ratio. The experimental results show that the accuracy and recall of the two algorithms have improved significantly, which indicates that the model training of the two algorithms tends to characterize the data features with a high proportion, that is, SVM and GBDT are not sensitive to the minority.
The accuracy of partial algorithms decreases when the data is increasing because the selection of the datasets is random. Besides, with the increase of data, the weight of learning will change, which also affects the accuracy.
Basically, the performance of ensemble algorithms generally outperforms single model algorithms in the interference discrimination problem, which indicates that ensemble algorithms are available for characterizing complex nonlinear relationships. Besides, the accuracy of partial algorithms decreases when the data is increasing because the selection of datasets is random. Besides, with the increase of data, the weight of learning will change, which also affects the accuracy.

4.4. Sensitivity of the Algorithms to IR

Typically, IR refers to the ratio of the majority to the minority in the training set. In this paper, IR represents the ratio of undisturbed samples to interfered samples in the training set. To verify the influence of IR on algorithms, the IR of the training set is set as 3:1, 5:1, 7:1, 9:1, and 11:1, respectively. The size of all training sets is 40,000. Meanwhile, the equivalent data is sampled per hour to form each training set.
As mentioned, the size of the test set is set to 20% of the number of the corresponding training set. The number of the interfered data and the normal data are the same in each test set. The equivalent data is sampled per hour to form the test set. Besides, there is no intersection between the training set and the test set, and the dataset is selected randomly. Similarly, accuracy and recall are applied to evaluate the algorithms.
The impact of IR on the algorithms is listed in Table 4. The accuracy results of single classification algorithms and ensemble algorithms are shown in Figure 9a. The recall results of two kinds of algorithms are shown in Figure 9b.
It is shown in Figure 9a that with the increase of the IR, the accuracy results of all algorithms decrease by degrees, which means that IR has an appreciable effect on the algorithms. When the IR is 3:1, the results among Bagging, Stacking, XGBoost, and RF are close. It means that when the value of IR is small, the ensemble algorithms are capable of achieving comparatively thorough learning of the dataset. However, with the increase of IR, the decline range of Bagging is smaller than the others, which validates that Bagging is able to learn the highly imbalanced dataset well.
With the increase of the IR, the accuracy results of Stacking, XGBoost, and RF are dropping obviously. When the IR is 11:1, the results of the three algorithms are close to the result of kNN. Moreover, similar results could be found in Figure 9b. The recall of kNN is even higher than that of XGBoost. It is reasonable that IR has a great impact on the ensemble algorithms, that is, the characteristics of the minority in highly imbalanced datasets are difficult to learn. Meanwhile, the reduction of the minority means that the characteristics of the minority will be more prominent, so kNN is easy to match the point at this time.
As mentioned before, NB is driven by a few samples, so the performance of NB changes little. The performance of AdaBoost is still not improved on the imbalanced dataset due to the weight distribution problem.
From the experimental results illustrated in Figure 10 and Figure 11, SVM and GBDT are not sensitive to the minority. However, it is observed that when the IR is 3:1, the accuracy of SVM is 50.99% and the recall of SVM is 2.33%. It means that SVM is able to be utilized to characterize the minority only when the IR is low enough, which further confirms that the learning ability of SVM for the imbalanced dataset is weak.

4.5. Robustness Analysis of the Algorithms

Data measurement failure caused by equipment power failure is unavoidable. In consequence, the abnormal data is included in our dataset considering the actual equipment conditions. The main forms of the abnormal data are the down tilt angle, equaling − 1 , when the antenna height is 0, and so forth. Some abnormal data is added to the training set to analyze the robustness of the algorithms.
We adopt the training set of Part C as the initial training set of the experiment. The IR of the training set is still 5:1. In the following, the abnormal data randomly replaces the same amount of data in the training set, and the replaced proportion is 1% and 5% of the training set, respectively.
The test set does not change in all experiments. About 1000 pieces of abnormal data are employed to form the test set. The equivalent abnormal data is sampled per hour to form the test set. In addition, there is no overlap between the training set and the test set. The accuracy is used for evaluating the robustness of the algorithms.
Table 5 shows the learning ability of the algorithms for abnormal data. The accuracy results of algorithms, which are trained by the 1% dataset, are shown in Figure 10. It can be seen that with the increase of the training data, the accuracy results of most algorithms are improving. The accuracy of XGBoost is higher than the others, which means that XGBoost could learn the characteristics of abnormal data well even if the number of data is small. Moreover, the performance of RF, kNN, and Bagging is also stationary.
The accuracy results of SVM, AdaBoost, NB, and Stacking are 63.63% when the training set contains 1% abnormal data. By analyzing the test set, we find that the data, which is not affected by the atmospheric duct, accounts for 63.63% of the training set. It means that the above four algorithms are not sensitive to samples when the number of samples is extremely low.
Figure 11 presents the robustness of the algorithms on the training set with 5% abnormal data. It is observed that the increase of the abnormal data from 1% to 5% improves the accuracy of the algorithms. Stacking outperforms other algorithms. In Figure 11, 40,000 pieces of training data achieve higher accuracy than that of 50,000 pieces, which indicates that the data characteristics contained in the randomly selected database have not been well learned by the algorithms. The accuracy difference between 40,000 and 50,000 data is about 1%, which indicates that the random data selection will cause fluctuations, but there is no large deviation,.
Compared to Figure 10 and Figure 11, it can be known that the increase of the abnormal data from 1% to 5% greatly improves the accuracy of kNN and Stacking, which means the two algorithms will be trained well when the number of the abnormal data reaches a certain level, but it also reflects that they are not sensitive to a few samples in a highly imbalanced dataset.
Moreover, AdaBoost is also greatly affected by the number of abnormal data, although the accuracy is not ideal. However, the increase of the abnormal data does not improve the accuracy of SVM and GBDT, which means the learning ability of the two algorithms is weak when the dataset is a highly imbalanced set and there are complex nonlinear relationships between the variables. Besides, with the increase of the abnormal data, the accuracy of NB changes slightly, which means that NB is sensitive to the abnormal data, that is, NB has ordinary learning ability for the highly imbalanced dataset.

4.6. Time Complexity

To analyze the algorithm efficiency, we list the time complexity of each algorithm, namely, the floating-point operations. To ensure comparison consistency, the time complexity is the result of running the code once in each algorithm. The time complexity and order of the algorithms are listed in Table 6 where n represents the number of inputs.
The time complexity is explained in detail. k denotes the dimension of a single sample characteristics. c indicates the number of categories. m represents the number of decision trees. d refers to the depth of the tree. | | x | | 0 means all non missing items in the training data. The order of Bagging and Stacking is related to the time complexity of base classifiers.
Specifically, the order of SVM is quadratic, which is unfriendly to the problem with considerable training data. The order of Bagging and Stacking depends on the selected base classifier, that is, when the order of the base classifier is low, the time complexity of Bagging and Stacking is acceptable. XGBoost adopts fractional data block parallelism, which enables the time complexity competitive.
To intuitively compare the complexity of the algorithm, we run the program in the configuration environment of Part B, and listed the test time in Table 6. Without loss of generality, each algorithm only compares the training time. The training set is selected from Part C, the data size is 40,000, and the IR is 5:1.
The time consumption of algorithms is shown in Table 6. It can be found that although the order of ensemble algorithms is generally higher than that of single model algorithms, its time consumption in solving the complex interference discrimination is still acceptable.

5. Conclusions

In this paper, a remote interference discrimination testbed with several promising AI algorithms was proposed to assist operators in identifying interference. The introduced framework for the testbed and the detailed design of the modules were presented. Furthermore, the testbed with 5,520,000 network-side data made a consistent comparison of nine AI algorithms. Numerical results illustrated that the ensemble algorithm had higher interference discrimination accuracy than the single model algorithm. Operators could select the algorithm with appropriate complexity to discriminate interference according to the conditions of hardware equipment. Considering the fluctuating accuracy of the algorithm, future work will consider optimizing the ability of the algorithm to learn data characteristics so that the algorithm can achieve stable performance. Moreover, the accuracy upper bound of remote interference discrimination deserves further exploration.

Author Contributions

Conceptualization, H.Z. and T.Z.; methodology, H.Z.; software, T.X.; validation, H.H.; writing—original draft preparation, H.Z.; writing—review and editing, T.Z.; supervision, T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Commission Foundation of Shanghai (Nos. 21511101400 and 22511100600).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoTInternet of Things
TDDTime-Division Duplex
GPGuard Period
kNNk-Nearest Neighbors
SVMSupport Vector Machine
NBNaive Bayes
RFRandom Forest
AdaBoostAdaptive Boosting
GBDTGradient Boosting Decision Tree
XGBoostXtreme Gradient Boosting
BaggingBootstrap Aggregating
StackingStacked Generalization
PEParabolic Equation
IRImbalance Ratio

References

  1. Chen, X.; Wu, C.; Chen, T.; Liu, Z.; Zhang, H.; Bennis, M.; Liu, H.; Ji, Y. Information Freshness-Aware Task Offloading in Air-Ground Integrated Edge Computing Systems. IEEE J. Sel. Areas Commun. 2022, 40, 243–258. [Google Scholar] [CrossRef]
  2. Cao, S.; Chen, X.; Zhang, X.; Chen, X. Improving the Tracking Accuracy of TDMA-Based Acoustic Indoor Positioning Systems Using a Novel Error Correction Method. IEEE Sens. J. 2022, 22, 5427–5436. [Google Scholar] [CrossRef]
  3. Xia, T.; Wang, M.M.; Zhang, J.; Wang, L. Maritime Internet of Things: Challenges and Solutions. IEEE Wirel. Commun. 2020, 27, 188–196. [Google Scholar] [CrossRef]
  4. Wagner, M.; Groll, H.; Dormiani, A.; Sathyanarayanan, V.; Mecklenbräuker, C.; Gerstoft, P. Phase Coherent EM Array Measurements in A Refractive Environment. IEEE Trans. Antennas Propag. 2021, 69, 6783–6796. [Google Scholar] [CrossRef]
  5. Zhang, H.; Zhou, T.; Xu, T.; Wang, Y.; Hu, H. Statistical Modeling of Evaporation Duct Channel for Maritime Broadband Communications. IEEE Trans. Veh. Technol. 2022, 71, 10228–10240. [Google Scholar] [CrossRef]
  6. Fang, X.; Feng, W.; Wei, T.; Chen, Y.; Ge, N.; Wang, C.X. 5G Embraces Satellites for 6G Ubiquitous IoT: Basic Models for Integrated Satellite Terrestrial Networks. IEEE Internet Things J. 2021, 8, 14399–14417. [Google Scholar] [CrossRef]
  7. Son, H.K.; Hong, H.J. Interference Analysis through Ducting on Korea’s LTE-TDD System from Japan’s WiMAX. In Proceedings of the 2014 International Conference on Information and Communication Technology Convergence (ICTC), Busan, Republic of Korea, 22–24 October 2014; pp. 802–805. [Google Scholar]
  8. Colussi, L.C.; Schiphorst, R.; Teinsma, H.W.M.; Witvliet, B.A.; Fleurke, S.R.; Bentum, M.J.; van Maanen, E.; Griffioen, J. Multiyear Trans-Horizon Radio Propagation Measurements at 3.5 GHz. IEEE Trans. Antennas Propag. 2018, 66, 884–896. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, Q.; Burkholder, R.J. Modeling And Measurement of Ducted EM Propagation over The Gulf Stream. In Proceedings of the 2019 IEEE International Symposium on Antennas and Propagation and USNC-URSI Radio Science Meeting, Atlanta, GA, USA, 7–12 July 2019; pp. 167–168. [Google Scholar]
  10. Zhang, H.; Zhou, T.; Xu, T.; Wang, Y.; Hu, H. FNN-Based Prediction of Wireless Channel with Atmospheric Duct. In Proceedings of the IEEE International Conference on Communications Workshops, Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar]
  11. 3rd Generation Partnership Project (3GPP). Technical Specification Group Radio Access Network; Study on Remote Interference Management for NR (Release 16); 3GPP TR 38.866 V16.1.0; Alliance for Telecommunications Industry Solutions: Washington, DC, USA, 2019. [Google Scholar]
  12. International Telecommunication Union (ITU). The Radio Refractive Index: Its Formula And Refractivity Data; ITU-R P.453-14; International Telecommunications Union—Radiocommunications Sector (ITU-R): Geneva, Switzerland, 2019. [Google Scholar]
  13. Wei, T.; Feng, W.; Chen, Y.; Wang, C.X.; Ge, N.; Lu, J. Hybrid Satellite-Terrestrial Communication Networks for The Maritime Internet of Things: Key Technologies, Opportunities, And Challenges. IEEE Internet Things J. 2021, 8, 8910–8934. [Google Scholar] [CrossRef]
  14. Xu, L.; Yardim, C.; Mukherjee, S.; Burkholder, R.J.; Wang, Q.; Fernando, H.J.S. Frequency Diversity in Electromagnetic Remote Sensing of Lower Atmospheric Refractivity. IEEE Trans. Antennas Propag. 2022, 70, 547–558. [Google Scholar] [CrossRef]
  15. Gilles, M.A.; Earls, C.; Bindel, D. A Subspace Pursuit Method to Infer Refractivity in the Marine Atmospheric Boundary Layer. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5606–5617. [Google Scholar] [CrossRef]
  16. Feng, G.; Huang, J.; Su, H. A New Ray Tracing Method Based on Piecewise Conformal Transformations. IEEE Trans. Microw. Theory Tech. 2022, 70, 2040–2052. [Google Scholar] [CrossRef]
  17. Dinc, E.; Akan, O.B. Channel Model for The Surface Ducts: Large-Scale Path-Loss, Delay Spread, And AOA. IEEE Trans. Antennas Propag. 2015, 63, 2728–2738. [Google Scholar] [CrossRef] [Green Version]
  18. Ozgun, O.; Sahin, V.; Erguden, M.E.; Apaydin, G.; Yilmaz, A.E.; Kuzuoglu, M.; Sevgi, L. PETOOL v2.0: Parabolic Equation Toolbox with evaporation duct models and real environment data. Comput. Phys. Commun. 2020, 256, 107454. [Google Scholar] [CrossRef]
  19. Huang, L.F.; Liu, C.G.; Wang, H.G.; Zhu, Q.L.; Zhang, L.J.; Han, J.; Wang, Q.N. Experimental Analysis of Atmospheric Ducts and Navigation Radar Over-the-Horizon Detection. Remote Sens. 2022, 14, 2588. [Google Scholar] [CrossRef]
  20. Wang, H.; Su, S.; Tang, H.; Jiao, L.; Li, Y. Atmospheric Duct Detection Using Wind Profiler Radar and RASS. J. Atmos. Ocean. Technol. 2019, 36, 557–565. [Google Scholar] [CrossRef]
  21. Zhou, T.; Sun, T.; Hu, H.; Xu, H.; Yang, Y.; Harjula, I.; Koucheryavy, Y. Analysis and Prediction of 100 km-Scale Atmospheric Duct Interference in TD-LTE Networks. J. Commun. Inf. Netw. 2017, 2, 66–80. [Google Scholar] [CrossRef] [Green Version]
  22. L’Hour, C.A.; Fabbro, V.; Chabory, A.; Sokoloff, J. 2-D Propagation Modeling in Inhomogeneous Refractive Atmosphere Based on Gaussian Beams Part I: Propagation Modeling. IEEE Trans. Antennas Propag. 2019, 67, 5477–5486. [Google Scholar] [CrossRef]
  23. Apaydin, G.; Sevgi, L. Matlab-Based Fem-Parabolic-Equation Tool for Path-Loss Calculations along Multi-Mixed-Terrain Paths. IEEE Antennas Propag. Mag. 2014, 56, 221–236. [Google Scholar] [CrossRef]
  24. Obaidat, M.A.; Obeidat, S.; Holst, J.; Al Hayajneh, A.; Brown, J. A Comprehensive and Systematic Survey on the Internet of Things: Security and Privacy Challenges, Security Frameworks, Enabling Technologies, Threats, Vulnerabilities and Countermeasures. Computers 2020, 9, 44. [Google Scholar] [CrossRef]
  25. Moreta, C.E.G.; Acosta, M.R.C.; Koo, I. Prediction of Digital Terrestrial Television Coverage Using Machine Learning Regression. IEEE Trans. Broadcast. 2019, 65, 702–712. [Google Scholar] [CrossRef]
  26. Sun, T.; Zhou, T.; Xu, H.; Yang, Y. A Random Forest-Based Prediction Method of Atmospheric Duct Interference in TD-LTE Networks. In Proceedings of the 2017 IEEE Globecom Workshops, Singapore, 4–8 December 2017; pp. 1–6. [Google Scholar]
  27. Gopi, S.P.; Magarini, M.; Alsamhi, S.H.; Shvetsov, A.V. Machine Learning-Assisted Adaptive Modulation for Optimized Drone-User Communication in B5G. Drones 2021, 5, 128. [Google Scholar] [CrossRef]
  28. Niu, J.; Wang, B.; Shu, L.; Duong, T.Q.; Chen, Y. ZIL: An Energy-Efficient Indoor Localization System Using ZigBee Radio to Detect WiFi Fingerprints. IEEE J. Sel. Areas Commun. 2015, 33, 1431–1442. [Google Scholar] [CrossRef] [Green Version]
  29. Luo, F.; Poslad, S.; Bodanese, E. Human Activity Detection and Coarse Localization Outdoors Using Micro-Doppler Signatures. IEEE Sens. J. 2019, 19, 8079–8094. [Google Scholar] [CrossRef]
  30. Peppes, N.; Daskalakis, E.; Alexakis, T.; Adamopoulou, E.; Demestichas, K. Performance of Machine Learning-Based Multi-Model Voting Ensemble Methods for Network Threat Detection in Agriculture 4.0. Sensors 2021, 21, 7475. [Google Scholar] [CrossRef]
  31. Kafai, M.; Eshghi, K. CROification: Accurate Kernel Classification with The Efficiency of Sparse Linear SVM. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 34–48. [Google Scholar] [CrossRef] [PubMed]
  32. Muñoz, E.C.; Pedraza, L.F.; Hernández, C.A. Machine Learning Techniques Based on Primary User Emulation Detection in Mobile Cognitive Radio Networks. Sensors 2022, 22, 4659. [Google Scholar] [CrossRef]
  33. Yang, B.; Lei, Y.; Yan, B. Distributed Multi-Human Location Algorithm Using Naive Bayes Classifier for A Binary Pyroelectric Infrared Sensor Tracking System. IEEE Sens. J. 2016, 16, 216–223. [Google Scholar] [CrossRef]
  34. Page, A.; Sagedy, C.; Smith, E.; Attaran, N.; Oates, T.; Mohsenin, T. A Flexible Multichannel EEG Feature Extractor And Classifier for Seizure Detection. IEEE Trans. Circuits Syst. II Express Briefs 2015, 62, 109–113. [Google Scholar] [CrossRef]
  35. Eldesouky, E.; Bekhit, M.; Fathalla, A.; Salah, A.; Ali, A. A Robust UWSN Handover Prediction System Using Ensemble Learning. Sensors 2021, 21, 5777. [Google Scholar] [CrossRef]
  36. Wu, D.; Jiang, Z.; Xie, X.; Wei, X.; Yu, W.; Li, R. LSTM Learning with Bayesian And Gaussian Processing for Anomaly Detection in Industrial IoT. IEEE Trans. Ind. Inform. 2020, 16, 5244–5253. [Google Scholar] [CrossRef] [Green Version]
  37. Oh, H. A YouTube Spam Comments Detection Scheme Using Cascaded Ensemble Machine Learning Model. IEEE Access 2021, 9, 144121–144128. [Google Scholar] [CrossRef]
  38. Alexandridis, A.; Chondrodima, E.; Giannopoulos, N.; Sarimveis, H. A Fast And Efficient Method for Training Categorical Radial Basis Function Networks. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2831–2836. [Google Scholar] [CrossRef] [PubMed]
  39. Ren, Z.A.; Chen, Y.P.; Liu, J.Y.; Ding, H.J.; Wang, Q. Implementation of Machine Learning in Quantum Key Distributions. IEEE Commun. Lett. 2021, 25, 940–944. [Google Scholar] [CrossRef]
  40. Bhowan, U.; Johnston, M.; Zhang, M.; Yao, X. Evolving Diverse Ensembles Using Genetic Programming for Classification with Unbalanced Data. IEEE Trans. Evol. Comput. 2013, 17, 368–386. [Google Scholar] [CrossRef] [Green Version]
  41. Salman, E.H.; Taher, M.A.; Hammadi, Y.I.; Mahmood, O.A.; Muthanna, A.; Koucheryavy, A. An Anomaly Intrusion Detection for High-Density Internet of Things Wireless Communication Network Based Deep Learning Algorithms. Sensors 2023, 23, 206. [Google Scholar] [CrossRef]
  42. Dutta, A.; Dasgupta, P. Ensemble Learning with Weak Classifiers for Fast And Reliable Unknown Terrain Classification Using Mobile Robots. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 2933–2944. [Google Scholar] [CrossRef]
  43. Lettich, F.; Lucchese, C.; Nardini, F.M.; Orlando, S.; Perego, R.; Tonellotto, N.; Venturini, R. Parallel Traversal of Large Ensembles of Decision Trees. IEEE Trans. Parallel Distrib. Syst. 2019, 30, 2075–2089. [Google Scholar] [CrossRef] [Green Version]
  44. Dhibi, K.; Fezai, R.; Mansouri, M.; Trabelsi, M.; Kouadri, A.; Bouzara, K.; Nounou, H.; Nounou, M. Reduced Kernel Random Forest Technique for Fault Detection And Classification in Grid-Tied PV Systems. IEEE J. Photovolt. 2020, 10, 1864–1871. [Google Scholar] [CrossRef]
  45. Alrowais, F.; Marzouk, R.; Nour, M.K.; Mohsen, H.; Hilal, A.M.; Yaseen, I.; Mohammed, G.P. Intelligent Intrusion Detection Using Arithmetic Optimization Enabled Density Based Clustering with Deep Learning. Electronics 2022, 11, 3541. [Google Scholar] [CrossRef]
  46. Quadrianto, N.; Ghahramani, Z. A Very Simple Safe-Bayesian Random Forest. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1297–1303. [Google Scholar] [CrossRef] [Green Version]
  47. Nguyen, T.T.; Pham, X.C.; Liew, A.W.C.; Pedrycz, W. Aggregation of Classifiers: A Justifiable Information Granularity Approach. IEEE Trans. Cybern. 2019, 49, 2168–2177. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Zhang, L.; Suganthan, P.N. Benchmarking Ensemble Classifiers with Novel Co-Trained Kernel Ridge Regression And Random Vector Functional Link Ensembles [Research Frontier]. IEEE Comput. Intell. Mag. 2017, 12, 61–72. [Google Scholar] [CrossRef]
  49. Khan, F.H.; Saadeh, W. An EEG-Based Hypnotic State Monitor for Patients During General Anesthesia. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 2021, 29, 950–961. [Google Scholar] [CrossRef]
  50. Iwendi, C.; Khan, S.; Anajemba, J.H.; Mittal, M.; Alenezi, M.; Alazab, M. The Use of Ensemble Models for Multiple Class and Binary Class Classification for Improving Intrusion Detection Systems. Sensors 2020, 20, 2559. [Google Scholar] [CrossRef]
  51. Sun, P.Z.; You, J.; Qiu, S.; Wu, E.Q.; Xiong, P.; Song, A.; Zhang, H.; Lu, T. AGV-Based Vehicle Transportation in Automated Container Terminals: A Survey. IEEE Trans. Intell. Transp. Syst. 2022, 24, 341–356. [Google Scholar] [CrossRef]
  52. Chen, K.; Guo, S. RASP-Boost: Confidential Boosting-Model Learning with Perturbed Data in The Cloud. IEEE Trans. Cloud Comput. 2018, 6, 584–597. [Google Scholar] [CrossRef]
  53. Liu, T.J.; Liu, K.H.; Lin, J.Y.; Lin, W.; Kuo, C.C.J. A ParaBoost Method to Image Quality Assessment. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 107–121. [Google Scholar] [CrossRef]
  54. Ji, X.; Yang, B.; Tang, Q. Acoustic Seabed Classification Based on Multibeam Echosounder Backscatter Data Using The PSO-BP-AdaBoost Algorithm: A Case Study from Jiaozhou Bay, China. IEEE J. Ocean. Eng. 2021, 46, 509–519. [Google Scholar] [CrossRef]
  55. Zhang, F.; Wang, Y.; Ni, J.; Zhou, Y.; Hu, W. SAR Target Small Sample Recognition Based on CNN Cascaded Features And AdaBoost Rotation Forest. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1008–1012. [Google Scholar] [CrossRef]
  56. Zhang, Y.; Wang, J.; Wang, X.; Xue, Y.; Song, J. Efficient Selection on Spatial Modulation Antennas: Learning or Boosting. IEEE Wirel. Commun. Lett. 2020, 9, 1249–1252. [Google Scholar] [CrossRef]
  57. Wu, W.; Xia, Y.; Jin, W. Predicting Bus Passenger Flow And Prioritizing Influential Factors Using Multi-Source Data: Scaled Stacking Gradient Boosting Decision Trees. IEEE Trans. Intell. Transp. Syst. 2021, 22, 2510–2523. [Google Scholar] [CrossRef]
  58. Zhang, W.; Wang, L.; Chen, J.; Xiao, W.; Bi, X. A Novel Gas Recognition And Concentration Detection Algorithm for Artificial Olfaction. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
  59. Xia, S.; Wang, G.; Chen, Z.; Duan, Y.; Liu, Q. Complete Random Forest Based Class Noise Filtering Learning for Improving The Generalizability of Classifiers. IEEE Trans. Knowl. Data Eng. 2019, 31, 2063–2078. [Google Scholar] [CrossRef]
  60. Singh, N.; Choe, S.; Punmiya, R.; Kaur, N. XGBLoc: XGBoost-Based Indoor Localization in Multi-Building Multi-Floor Environments. Sensors 2022, 22, 6629. [Google Scholar] [CrossRef] [PubMed]
  61. Wu, D.; Lin, C.T.; Huang, J.; Zeng, Z. On The Functional Equivalence of TSK Fuzzy Systems to Neural Networks, Mixture of Experts, CART, And Stacking Ensemble Regression. IEEE Trans. Fuzzy Syst. 2020, 28, 2570–2580. [Google Scholar] [CrossRef] [Green Version]
  62. Zhu, H.; Li, Y.; Li, R.; Li, J.; You, Z.; Song, H. SEDMDroid: An Enhanced Stacking Ensemble Framework for Android Malware Detection. IEEE Trans. Netw. Sci. Eng. 2021, 8, 984–994. [Google Scholar] [CrossRef]
  63. Saha, S.; Moorthi, S.; Wu, X.; Wang, J.; Nadiga, S.; Tripp, P.; Behringer, D.; Hou, Y.T.; Chuang, H.Y.; Iredell, M.; et al. The NCEP Climate Forecast System Version 2. J. Clim. 2014, 27, 2185–2208. [Google Scholar] [CrossRef]
Figure 1. Remote interference in TDD system, in which GP is the guard period and PTS is the pilot time slot.
Figure 1. Remote interference in TDD system, in which GP is the guard period and PTS is the pilot time slot.
Sensors 23 02264 g001
Figure 2. The framework of the testbed.
Figure 2. The framework of the testbed.
Sensors 23 02264 g002
Figure 3. SVM hyperplane for discrimination.
Figure 3. SVM hyperplane for discrimination.
Sensors 23 02264 g003
Figure 4. Structure of the random forest classifier.
Figure 4. Structure of the random forest classifier.
Sensors 23 02264 g004
Figure 5. Structure of Bagging classifier.
Figure 5. Structure of Bagging classifier.
Sensors 23 02264 g005
Figure 6. Schematic of a stacking classifier framework.
Figure 6. Schematic of a stacking classifier framework.
Sensors 23 02264 g006
Figure 7. Number of interfered base stations.
Figure 7. Number of interfered base stations.
Sensors 23 02264 g007
Figure 8. Accuracy and recall results of single model algorithms and ensemble algorithms with different data sizes. (a) Accuracy results. (b) Recall results.
Figure 8. Accuracy and recall results of single model algorithms and ensemble algorithms with different data sizes. (a) Accuracy results. (b) Recall results.
Sensors 23 02264 g008
Figure 9. Accuracy and recall results of single model algorithms and ensemble algorithms with different IR. (a) Accuracy results. (b) Recall results.
Figure 9. Accuracy and recall results of single model algorithms and ensemble algorithms with different IR. (a) Accuracy results. (b) Recall results.
Sensors 23 02264 g009
Figure 10. Accuracy results of algorithms in the training set that contains 1% abnormal data.
Figure 10. Accuracy results of algorithms in the training set that contains 1% abnormal data.
Sensors 23 02264 g010
Figure 11. Accuracy results of algorithms in the training set that contains 5% abnormal data.
Figure 11. Accuracy results of algorithms in the training set that contains 5% abnormal data.
Sensors 23 02264 g011
Table 1. Hardware and software configurations.
Table 1. Hardware and software configurations.
DesignationConfiguration
Corei5-4210 H 2.90 GHz
Operating systemWindows 10
Random-access memory12.0 GB
Python3.7
Tensorflow2.0.0
Table 2. Settings of the key parameters in the algorithms.
Table 2. Settings of the key parameters in the algorithms.
CategoryAlgorithmsParametersValue
Single
model
algorithms
kNNNumber of neighbors1
SVMKernelLinear/Radial basis function
Maximum number of iterations100
NBTypeGaussian/Bernoulli/Complement
Ensemble
algorithms
RFNumber of trees in the forest100
BaggingNumber of base estimators in the ensemble100
BoostingAdaBoostMaximum number of estimators500
Learning rate0.01
GBDTNumber of boosting stages to perform100
Learning rate0.01
XGBoostNumber of decision trees100
Learning rate0.1
StackingEstimatorsLr/rf/kNN/cart/svc/bayes
Final estimatorLogisticRegression
Table 3. Results of changing the data size of the algorithms.
Table 3. Results of changing the data size of the algorithms.
IR5:1
Size of Training Data20,00030,00040,00050,00060,000
IndicatorsAccRecallAccRecallAccRecallAccRecallAccRecall
Single
model
algorithms
kNN61.9336.4663.6538.1464.1741.2165.8343.4965.5242.67
SVM50.000.0050.020.4650.010.3350.000.0050.000.00
NB54.7417.5854.7617.1654.6918.1355.1418.1954.7417.73
Ensemble
algorithms
RF70.1843.2371.2745.6172.3848.0773.8350.7374.0050.92
Bagging76.0359.0478.3262.1579.6764.8479.6765.6181.4467.93
BoostingAdaBoost53.757.8554.388.8954.659.8854.509.3354.129.23
GBDT50.000.0050.000.0050.000.0050.000.0050.000.00
XGBoost69.5841.8970.5643.3171.9846.7372.4947.9872.8848.14
Stacking71.4945.9972.8247.9574.6852.3476.9656.7877.2857.38
Table 4. Results of changing the IR on the algorithms.
Table 4. Results of changing the IR on the algorithms.
Size of Training Data40,000
IR3:15:17:19:111:1
IndicatorsAccRecallAccRecallAccRecallAccRecallAccRecall
Single
model
algorithms
kNN67.1051.5764.1741.2161.9234.2661.5930.9860.2627.55
SVM50.992.3350.010.3350.000.0050.000.0050.000.00
NB54.7121.0454.6918.1354.5215.9054.3014.9654.0913.81
Ensemble
algorithms
RF77.5160.2472.3848.0769.3540.9965.9133.2963.6128.59
Bagging82.3772.8979.6764.8476.3757.1673.6851.2071.3746.33
BoostingAdaBoost54.689.7154.659.8852.395.0651.984.0451.903.82
GBDT50.000.0050.000.0050.000.0050.000.0050.000.00
XGBoost77.8560.6971.9846.7367.4936.3764.0429.1161.1923.08
Stacking81.0467.7074.6852.3470.4542.6366.2633.7964.2929.81
Table 5. Results of the algorithms on abnormal data.
Table 5. Results of the algorithms on abnormal data.
IndicatorAccuracy
Size of Training Data20,00030,00040,00050,00060,000
Proportion of Abnormal Data1%5%1%5%1%5%1%5%1%5%
Single
model
algorithms
kNN77.7691.5080.3393.8784.0997.2386.8593.9786.9596.83
SVM63.6363.6363.6363.6363.6363.6363.6363.6363.6363.63
NB63.6363.8363.4363.7363.5463.9363.5463.4363.4363.63
Ensemble
algorithms
RF88.7395.2590.3195.5590.6196.6491.4095.2592.2995.75
Bagging89.3296.3490.5196.2491.1096.7392.1995.0692.6896.65
BoostingAdaBoost63.6383.5963.6383.7963.6383.7063.6380.8363.6380.73
GBDT63.6363.6363.6363.6363.6363.6363.6363.6363.6363.63
XGBoost89.4296.3490.9096.5491.6997.0392.4995.8492.8897.13
Stacking63.6396.4463.6396.7363.6397.9263.6396.5463.6397.82
Table 6. Time complexity of the algorithms.
Table 6. Time complexity of the algorithms.
AlgorithmsTime ComplexityOrderTest
kNN O ( k n ) O ( n ) 1.49 s
SVM O ( n 2 ) O ( n 2 ) 8.20 s
NB O ( c k n ) O ( n ) 1.02 s
RF O ( k d m n l o g n ) O ( n l o g n ) 3.98 s
Bagging O ( B a s e ) O ( B a s e ) 11.52 s
AdaBoost O ( k n l o g n ) O ( n l o g n ) 3.49 s
GBDT O ( k d n l o g n ) O ( n l o g n ) 2.46 s
XGBoost O ( m d | | x | | 0 + | | x | | 0 l o g n ) O ( l o g n ) 1.79 s
Stacking O ( B a s e ) O ( B a s e ) 47.78 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Zhou, T.; Xu, T.; Hu, H. Remote Interference Discrimination Testbed Employing AI Ensemble Algorithms for 6G TDD Networks. Sensors 2023, 23, 2264. https://doi.org/10.3390/s23042264

AMA Style

Zhang H, Zhou T, Xu T, Hu H. Remote Interference Discrimination Testbed Employing AI Ensemble Algorithms for 6G TDD Networks. Sensors. 2023; 23(4):2264. https://doi.org/10.3390/s23042264

Chicago/Turabian Style

Zhang, Hanzhong, Ting Zhou, Tianheng Xu, and Honglin Hu. 2023. "Remote Interference Discrimination Testbed Employing AI Ensemble Algorithms for 6G TDD Networks" Sensors 23, no. 4: 2264. https://doi.org/10.3390/s23042264

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop