Next Article in Journal
Enhanced Communications on Satellite-Based IoT Systems to Support Maritime Transportation Services
Previous Article in Journal
A Centrifugal Pump Fault Diagnosis Framework Based on Supervised Contrastive Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Modulation Classification for MASK, MPSK, and MQAM Signals Based on Hierarchical Self-Organizing Map

College of Computer, National University of Defense Technology, Changsha 410073, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(17), 6449; https://doi.org/10.3390/s22176449
Submission received: 12 July 2022 / Revised: 23 August 2022 / Accepted: 24 August 2022 / Published: 26 August 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Automatic modulation classification (AMC) plays a fundamental role in common communication systems. Existing clustering models typically handle fewer modulation types with lower classification accuracies and more computational resources. This paper proposes a hierarchical self-organizing map (SOM) based on a feature space composed of high-order cumulants (HOC) and amplitude moment features. This SOM with two stacked layers can identify intrinsic differences among samples in the feature space without the need to set thresholds. This model can roughly cluster the multiple amplitude-shift keying (MASK), multiple phase-shift keying (MPSK), and multiple quadrature amplitude keying (MQAM) samples in the root layer and then finely distinguish the samples with different orders in the leaf layers. We creatively implement a discrete transformation method based on modified activation functions. This method causes MQAM samples to cluster in the leaf layer with more distinct boundaries between clusters and higher classification accuracies. The simulation results demonstrate the superior performance of the proposed hierarchical SOM on AMC problems when compared with other clustering models. Our proposed method can manage more categories of modulation signals and obtain higher classification accuracies while using fewer computational resources.

1. Introduction

Automatic modulation classification (AMC) refers to the practical and accurate identification of modulated signals in the case of insufficient prior information [1]. When receiving over-the-air signals in a complicated environment, it is a priority to recognize the modulation types of signals before decoding the received signal. In both the civil and military areas, AMC plays an important role in complex wireless non-cooperative communication environments [2]. The methods employed to solve AMC problems can largely be divided into two categories: one is based on maximum likelihood theory and the other is feature based [3]. The maximum likelihood methods, which incorporate decision theory and hypothesis testing theory, achieve better performance; however, they consume a large amount of computational resources and require a great deal of prior knowledge [4]. By contrast, feature-based methods typically extract features in the pre-processing stage and apply some neural networks as optional classifiers to recognize diverse modulation types [5]. The extracted features can describe the intrinsic characteristics of the signals and reflect the differences between them. Feature-based methods are robust to noise and benefit from decent explainability. Moreover, the whole calculation process for feature-based methods is relatively more concise than that for the maximum likelihood methods. Accordingly, feature-based methods are commonly and effectively applied to solve AMC problems and are widely used in different environments [6].
Multiple amplitude-shift keying (MASK), multiple phase-shift keying (MPSK), and multiple quadrature amplitude keying (MQAM) are commonly used in communication systems for AMC. MASK often appears in scenarios with low carrier wave frequency, such as radio frequency (RF) applications and industrial networks. This modulation method is easy to realize and has low computational resource requirements. MPSK is a digital modulation method with high spectrum utilization and strong anti-interference properties for noise on amplitude, which is used in various communication systems such as optical satellite and cellular phone networks [7]. This method is adopted to increase signal bandwidth and reduce incorrect transmissions. In addition to MASK and MPSK, MQAM modulation has been widely used in satellite communications and microwave communications with modulation of amplitude and phase due to its higher spectral efficiency [8]. Nowadays, applying AMC techniques to MASK, MPSK, and MQAM signals is a very useful method of demodulating signals and obtaining the information carried therein across many common communication systems [9]. This technology can also judge the source site of the received signals, and manage the communication resources by prohibiting the unauthorized communication sender.
There are different kinds of features used for AMC, including wavelet transform, cyclic spectrum, and high-order cumulants (HOC) [10]. Wavelet transform is a multi-resolution method of time and frequency analysis which can decompose modulated signals on different scales. It can represent the details of various signals and recognize different types of digital modulation. However, wavelet transform requires too many attempts to search for the appropriate scale and translation factors. Therefore, it is not suitable to recognize modulated signals with little prior knowledge. In addition, cyclic spectrum features represent the correlation characteristics of digital signals and achieve decent anti-noise performance [11]. Cyclic spectrum features also have sparsity characteristics and highlight the position of peak points. Some researchers use different cyclic spectrum peak values to distinguish between various modulated digital signals [12]. Furthermore, cyclostationarity-based features are applied in [13] to recognize MQAM signals. However, these features are obtained via the Fourier series and include several variables; thus, their computational resource consumption is far higher when compared with HOC features. HOC can alleviate the side effects of the additive white Gaussian noise (AWGN) and effectively extract the characteristics of the original signals. The HOC features contain high-order information for use in distinguishing between different modulation types [14]. The different theoretical values of HOC make it feasible to recognize various modulated signals.
Common AMC classifiers mainly include support vector machine (SVM), subtractive clustering, and self-organizing map (SOM). Considering that the received signals are typically derived from non-cooperative systems, it is challenging to identify the modulation types of signals without adequate prior knowledge; this situation is exacerbated given that the received signals are always affected by various noise and interference. Because the extracted features usually have high dimensions, linear models such as SVM perform poorly when processing several dimensions of data and classifying various categories of samples [15]. Notably, however, the clustering method can overcome the side effects of the SVM approaches due to its ability to manage high-dimensional data in a nonlinear manner. Subtractive clustering can classify different types of modulated signals without setting the number of clusters ahead of time; however, the process of carefully adjusting the acceptance and rejection thresholds of the cluster density is complex and intricate [16]. Differently, SOM can learn the topological structure of the input samples and the distribution of the sample characteristics. It can also identify the intrinsic differences between signals and work without needing to set fine-tuned thresholds. Kaur et al. [17] utilize SOM to generate clusters from wavelet feature vectors. Zhou et al. [18] add an inherent partition mechanism to make the SOM neural network more suitable for recognizing modulation types and demodulating digital signals. Xu et al. [19] improve the SOM algorithm by adjusting the learning rate and adopting the neighborhood function. The improved model achieves better performance than both subtractive clustering and the SVM model. SOM is able to cluster modulated signals with clear boundaries and few computation resources.
SOM was first proposed by Kohonen [20] to imitate the human brain’s specific responses to external information [21]. As an unsupervised learning model, SOM is a type of network that can respond selectively to input data. In more detail, SOM is able to construct a topology between input features and output spaces. Meanwhile, it is also a self-learning network composed of fully connected neuron arrays without numerous training datasets and tags. SOM is widely applied in the fields of industry, finance, natural sciences, and linguistics [22]. Melin et al. [23] use SOM to analyze the ways in which similar countries fought the COVID-19 pandemic together and accordingly propose corresponding strategies. In [24], SOM dynamically switches between supervised and unsupervised learning during the training according to the availability of the class tags in the dataset. Hierarchical SOM [25] is an improved form of SOM that contains several SOM layers in a hierarchical structure. It can present an automated visualization scheme and produce better partitioning by using an effective knowledge representation method [26].
To present obvious clustering results and increase classification accuracies, the proposed hierarchical SOM model roughly clusters samples in the root layer and generates other neurons in the leaf layers to cluster samples of different modulation orders. HOC and amplitude moments are applied as effective features to describe the intrinsic differences between MASK, MPSK, and MQAM digital signals and distinguish these signals in the proposed two-layer SOM model. Moreover, a discrete transformation method based on modified activation functions is used to create obvious clusters in the leaf layer for MQAM signals. The proposed hierarchical SOM model is found to obtain higher classification accuracies and consume fewer computational resources when compared with some other common classifiers in the following parts.
The contributions of this paper are as follows:
  • This paper proposes a method of generating leaf layers based on the root layer in a hierarchical SOM structure for AMC problems. This model can cluster the normalized features of different modulation types with diverse orders roughly in the root layer and finely in the leaf layers.
  • This paper presents the discrete transformation based on modified activation functions for the features of MQAM samples. This kind of novel transformation method can create clear boundaries of clusters in the leaf layer and produce higher classification accuracies.
  • This paper compares the hierarchical SOM model for automatic modulation classification with other common models. The proposed model can recognize more modulation types with higher classification accuracies and fewer computational resources.
The remainder of this paper is organized as follows. Section 2 introduces the selected features as the input of the SOM network. Section 3 provides the structure and learning procedures of the hierarchical SOM model. Section 4 discusses the discrete transformation based on modified activation functions for QAM features. The signal simulation method and experimental results are presented in Section 5, followed by a discussion and a conclusion.

2. Construction of Feature Space

At the receiving part, the down-converted and sampled signal can be expressed as follows:
r ( n ) = a e j 2 π f 0 n + θ 0 s ( n ) + w ( n )
where a is the attenuation factor, n is the time index of the baseband signal, f 0 is the signal frequency offset, θ 0 is the phase offset, s ( n ) is the transmitted symbol coming from M types of modulation signals, and w ( n ) is the complex AWGN with constant power.
HOC features have different characteristics when compared to the signal spectrum and the wavelet transform. Several cumulants have recently been proposed as features in the literature for addressing the AMC problem [27]. Cumulant-based methods have been widely used to obtain superior classification accuracy, and almost all previous works of this kind have considered the calculation of different combinations of cumulant features that are feasible for classifiers. This process aims to extract the features that can obtain high resolution and strong robustness at a low signal-to-noise ratio (SNR) [28]. The following expressions are used to calculate HOC based on signal amplitude moments.
The p-order mixing moment of the random process r ( t ) is
M p q = E r ( t ) p q r ( t ) q
where r ( t ) is the conjugate complex value of r ( t ) , while E [ · ] represents the mathematical expectation. The high-order cumulant of a zero-mean k-order stationary random process r ( t ) is defined as
C k r τ 1 , τ 2 , , τ k 1 = Cum r ( t ) , r t + τ 1 , , r t + τ k 1
where Cum · refers to the high-order cumulant of r ( t ) and τ is the time delay of r ( t ) . If r ( t ) is a Gaussian process, the value of mixing moments with more than two orders will be equal to zero. Therefore, the cumulants with three or more orders are extremely well suited to suppress Gaussian noise [14]. A Gaussian process is expressed in terms of HOC below.
C k r = Cum ( τ 1 , τ 2 , , τ k 1 ) 0 , k > 2
As a result, these features are robust to interference from Gaussian noise. Moreover, the absolute values of these features are independent of initial phases, which means that they are robust to interference from different initial phases. The theoretical values of cumulants can be obtained by certain formulas. With reference to some related studies [29,30,31], we can determine that the C 40 and C 42 cumulants of MASK, MPSK, and MQAM are different from each other; thus, they can serve as effective features for classifiers. The formulas for the C 40 and C 42 cumulants of various orders are as follows.
C 40 = Cum x , x , x , x = M 40 3 M 20 2
C 42 = Cum x , x , x , x = M 42 2 M 21 2 M 20 2
This paper constructs a feature space based on the HOC features outlined above to cluster feature samples. The first feature in this space is | C 42 | . The theoretical | C 42 | values of MPSK signals are greater than those of MQAM signals, meaning that it is feasible to cluster MPSK or MQAM samples in the root layer. The second feature in the feature space is | C 40 | . The PSK signals with various orders differ distinctly in terms of their | C 40 | values, meaning that PSK samples are able to be clustered in the leaf layer based on the | C 40 | feature.
However, the differences between 16QAM and 64QAM samples are difficult to distinguish. Accordingly, this paper applies an amplitude moment feature called μ 42 to distinguish different orders of QAM samples. μ 42 , which is also called the kurtosis of instantaneous amplitude, can describe the degree of compactness for the received signals: more specifically, signals with bigger compact amplitude values often have larger μ 42 values. The theoretical μ 42 value of 16QAM is 2.3590, while that of 64QAM reaches 2.5025 [32]. The differences in the μ 42 values of two kinds of QAM signals are accordingly obvious. The expressions of μ 42 are defined below.
μ 42 = E A c 4 ( n ) E 2 A c 2 ( n ) A c ( n ) = A ( n ) m a 1 , m a = 1 N n = 1 N A ( n )
In the above expressions, A c ( n ) is the central normalized instantaneous amplitude of the received signal r ( n ) , A ( n ) is the instantaneous amplitude of the received signal r ( n ) , and m a is the mean amplitude of one signal sample. A c can reflect the mean amplitude of the fluctuation signals without the side effects of direct current components.
The theoretical values of these features are in Table 1. Here, E is the energy value of the received signal; moreover, we use the μ 42 of MQAM signals rather than MPSK signals to cluster samples in the leaf layer. These three features can be effectively used to distinguish between all specified modulation types. The number of samples generated for each modulation type is S and the total number of samples is 7 S . The constructed feature space is a three-dimensional space containing several vectors. The feature of the i-th sample is expressed as X i in the equation below.
X i = [ C 42 , C 40 , μ 42 ] i i = 1 , 2 , , 7 S
It should be noted here that the performance of the clustering model is based on the construction of features. To avoid interference from the energy of different modulated signals, it is necessary to apply a normalization method to the feature space. This paper selects the Min-Max scaling method to maintain the relative relationship of the original space. Every feature sample is scaled independently based on the same feature. The expression for the Min-Max normalization method is presented below.
X ^ i j = X i j M i n ( X i j ) M a x ( X i j ) M i n ( X i j ) j = 1 , i [ 1 , 7 S ] j = 2 , i [ 3 S + 1 , 5 S ] j = 3 , i [ 5 S + 1 , 7 S ]
where i indicates the index of the samples, while j refers to the j-th feature of the original space. M a x ( X i j ) is the maximum value of the j-th feature when the indexes are within the specified range and M i n ( X i j ) is the minimum value. This normalization method attempts to regularize samples so that every sample can be transformed into a unit norm for each feature, which is expressed as X ^ i j . The values of X ^ i j range from zero to one. The feature samples are the input of the proposed model.

3. Structure and Algorithm of the Proposed Hierarchical SOM Model

SOM is an unsupervised learning system in which a group of connected neurons iteratively adapt to the distribution of input samples in the feature space. The remarkable characteristic of this model is that it maps similar input vectors to neighboring nodes in the output space. In essence, this mapping is a kind of reproduction of the original data, because it preserves the basic structure of the input space; it can learn the topological structure of the feature space consisting of the input samples in order to cluster similar samples near the mapping neuron in the output layer [33]. SOM is a vector-to-coding model that provides topological mapping to transform the input vectors into a topology with a specific pattern. Discrete mapping is carried out based on input feature samples with specified dimensions. This mapping is a self-adaptive process that progresses in a topological and orderly manner.
The proposed hierarchical SOM model is able to cluster numerous input vectors by finding a smaller number of prototypes to provide a good approximation of the original input space. The theoretical foundation of the hierarchical SOM model is rooted in vector quantization theory, the motivation for which is dimensionality control or data compression. Thus, the hierarchical SOM model can approximate the original space with different categories and use fewer computational resources in the process.

3.1. Hierarchical Structure of the SOM Model

SOM simulates the connections among neurons in the human brain to achieve information processing according to biodynamic principles. Based on the conventional SOM structure with only one competition layer, the hierarchical SOM structure is proposed here to overcome the drawbacks of the former, which include high computational consumption and low precision. This hierarchical structure can cluster the input vectors more finely based on the output in the root layer. For each neuron in the leaf layer of the hierarchical SOM, the input data are the vectors connected with the winning neuron in the root competition layer [34]. Different features of the vectors are applied to cluster the samples in the leaf layer. These input data mapped to the same winning neuron usually have similar values or topological characteristics, enabling similar data to be clustered in the new layer. The structure of the hierarchical SOM with two layers is presented in Figure 1.
The neurons in the leaf competition layer connect with all winning neurons in the root competition layer. The flexible and layering hierarchical SOM can produce more detailed clustering results than the original SOM. The hierarchical SOM works in the data clustering phase to finely cluster input samples with their representative features and pattern. Each SOM layer uses one of three features to increase the classification efficiency [35]. The goal of hierarchical SOM is to transform a classification problem with large categories into multiple small problems with few categories to improve classification accuracy. This hierarchical model is based on the idea of the dividing and ruling technique. To cluster different categories in a hierarchical SOM structure, we cluster the large categories with the root SOM layer and small categories of different modulation orders with the leaf SOM layer.
A significant feature of a hierarchical SOM network is that it retains the topological structure of the input space so that the adjacent sample data in the input space will be mapped to adjacent neurons. We calculate the position of the winning node corresponding to each input sample; moreover, the neurons in the output plane will reflect the number of categories with clustering results. We can directly obtain the recognition result through the category tag of the highest frequency within the neurons of the output layer. Another valid choice would be to use the frequency of winning neurons to estimate the probability density of each category. When a new sample has a certain possibility of falling into a specific neuron on the output plane, we can determine the recognition result accordingly. The clustering results in the output layer are derived from the hierarchical mapping.

3.2. Algorithm of the Hierarchical SOM Model

The proposed hierarchical SOM neural network mapping method has five main processes: initialization for weight vectors, competitive learning for the winning neuron, cooperation based on topological neighborhood function, adjustment for weight vectors, and the generation of new layers. It gradually optimizes the network based on competition among neurons and maintains the topology of the input space through the neighborhood function. In a departure from the existing literature on classical SOM models [36], the hierarchical SOM model also adds a layering procedure to generate a new mapping layer for accurate classification. The algorithm of the hierarchical model is shown in Algorithm 1.
Algorithm 1 Algorithm of the hierarchical SOM model
Require: Input the feature of samples x i ( i = 1 , 2 , , M )
Ensure: Output clustering results in the root and leaf layers
  1: Initialize weight vectors w j ( j = 1 , 2 , , N ) randomly
  2: for t = 0 to T 1  do
  3:    Find the winning neuron j = arg min x i w j
  4:    Define the topological neighborhood function near the winning neuron N j , k = exp ( d j , k 2 2 σ 2 ) ( k = 1 , 2 , . . . , P )
  5:    Adjust the weights of neurons in the topological neighborhood w k ( t + 1 ) = w k ( t ) + η ( t ) ( x i w j ) N j , k
  6:    if  q e = 1 M i = 1 M x i w j < τ  then
  7:      Generate the leaf layer and break
  8:    end if
  9:  end for
  10: Obtain the clustering results in the root layer
  11: Re-execute line 1 to line 5 for the leaf layers
  12: Obtain the clustering results in the leaf layers
We first need to set up the size of the SOM layer and initialize weight vectors with small values. In normal cases, it is necessary to consider that the ideal distribution of the initial weights roughly approximates the distribution of the input sample. Random initialization uses random numbers in a small range to initialize the original weight vectors and is suitable in cases where little or no prior knowledge of the input data is available. This initialization method starts the weight vectors from an arbitrary initial state [22]. The advantage of this approach is that the output network nodes are similar to the topology of the input data at the beginning and converge to a fixed topology during training. In most datasets with nonlinear characteristics and an unbalanced distribution of values, random initiation performs better.
The competition process aims to find the winning neuron in the competition layer whose connected weight vector is the most similar to the input vector x i ( i = 1 , 2 , . . . , M ) , where M is the whole number. Moreover, the weight vectors are w j ( j = 1 , 2 , . . . , N ) , where N is the number of neurons in the output competition layer. The discriminant function of competition works by finding the minimum Euclidean distance between the input vector x i and the weight vector w j of each neuron in the competition layer [36], which is as follows.
j = arg min j x i w j , i = 1 , 2 , , M
where each input vector x i in an SOM neural network has its winning neuron, which is the winner of the competition among weight neurons. If several weight vectors fulfill the discriminant function, the vector with the smallest subscript is the winner. The weight neuron that satisfies the discriminance and subscript conditions is referred to as the winning neuron.
When a winning neuron is active, the nearest neighbor nodes tend to be more excited than more distant neighbor nodes. A topological neighborhood function N j , k ( j = 1 , 2 , . . . , N ; k = 1 , 2 , . . . , P ) represents the influence area of the winning neuron j on the cooperative neuron k and the number of cooperative neurons is P. The winning neuron of the SOM neural network affects other neurons around it from near to far, and the effect weakens gradually from neural excitement to neural inhibition [37]. Therefore, the learning algorithm of the SOM neural network adjusts both the weight vector of the winning neuron and the weight vector of the surrounding neurons to varying degrees under the influence of the topological neighborhood function. The common neighborhood topology takes the form of a square, specifically four neurons in a square surrounding the winner neurons. N j , k is the topological neighborhood function centered on the winning neuron j and contains a group of cooperative neurons. One of these cooperative neurons is the neuron k. d j , k is the lateral distance between the winning neuron j and the excitatory neuron k. The topological neighborhood takes the maximum value when d j , k is both equal to zero and symmetrical about zero. The amplitude value of the neighborhood function N j , k decreases gradually with the increase of the lateral distance d j , k . The expression of the common Gaussian topological neighborhood function is as below.
N j , k = exp ( d j , k 2 2 σ 2 )
where σ is the shape parameter that controls the width of the function peak. Specifically, it indicates the degree of participation of excitatory neurons near the winning neuron in the learning process. In practical use, the positive values of the Gaussian function can bring the weight vectors in the neighborhood closer to the winning neuron.
After determining the topological neighborhood, SOM performs an adjustment process through which the output nodes self-organize to establish a regular feature mapping. Not only will the winning neuron cause the weights to update, but its neighbors will also update their weights [38]. The adjustment process strengthens the response of the winning neuron to a subsequent similar input vector. This adaptive process provides accurate statistical quantification of the input vectors. The adjustment process will end if the number of iterations reaches the maximum value T for convergence. The updating formula of the weight vectors in the neighborhood area, including the weight vector of the winning neuron, for the t-th step is as follows.
w k ( t + 1 ) = w k ( t ) + η ( t ) x i w j N j , k
where η ( t ) is the learning rate and N j , k ( t ) is the topological neighborhood function for the t-th step. Setting the original learning rate as η ( 0 ) , the iteration expression of η ( t ) is as below.
η ( t ) = η ( 0 ) 1 1 + 2 t / T , t = 1 , 2 , . . . , T
where T is the maximum number of iterations and η ( t ) decreases as t gradually increases. The decrease in the learning rate causes the SOM model to converge faster. Similarly, the shape parameter σ also becomes smaller with a larger number of iterations. The expression of the shape parameter for the t-th iteration, which is called σ ( t ) , is shown below.
σ ( t ) = σ ( 0 ) 1 1 + 2 t / T , t = 1 , 2 , . . . , T
The SOM model applies the quantization error to generate the new layers. The quantization error is the average distance between every input sample x i and its best matching unit w j . The new layer grows from the root SOM layer according to the quantization error in the original layer if the quantization error q e for the input samples in the competition layer fulfills the following network layering condition:
q e = 1 M i = 1 M x i w j < τ , t = 1 , 2 , , T
where M is the number of input vectors, j is the index of the best matching unit in the competition layer, | | . | | denotes the Euclidean distance between the input vector and the weight vector that connects the input vector and neuron in the competition layer, τ is the parameter that controls the generation of the leaf layer, and T is the maximum number of iterations. q e is able to measure the dispersion of all data mapped to a specific winning neuron and control the mapping procedure of the hierarchical SOM network. When a training iteration of the root layer ends, the algorithm will check whether all neurons on this layer meet the network layering condition. For any neuron j that fulfills this condition, a new layer is generated based on the root layer. For the neurons in the leaf layer of the hierarchical network, the input data are the sub-sequence of the original data that are mapped from the corresponding neurons in the root layer. Each layer in the model has a different size depending on the value distribution of the input features. When the new layer is generated from the root layer, the model returns to the initialization step for training with the original vectors like the root layer. If the number of iterations for neurons in the new layer eventually reaches the given iteration times, the proposed hierarchical SOM algorithm for clustering samples will terminate.

4. Discrete Transformation Based on Modified Activation Functions for QAM Features

As noted in Table 1, the differences between the theoretical values of 16QAM and 64QAM are almost imperceptible. To enhance the differentiation of the two QAM digital samples in the feature space, it is practical to apply a discrete transformation based on modified activation functions before clustering the specified features within the leaf layer. In fact, this kind of function makes the output vectors more discrete, meaning that the SOM network can distinctly cluster the output vectors with different orders of modulation. In this paper, the modified activation functions in the neural networks can complete the discrete transformation to map the original data into the specified interval ranges from zero to one. The discretized operation has no effects on the size of the SOM network and decreases the number of iterations required for convergence when training the leaf SOM layer. A number of activation functions, such as the arctangent function, can be employed to achieve discrete transformation. Moreover, the logistic function and hard-logistic function can also be modified to improve discrete transformation. We use s c a l e ( s c a l e > 0 ) to control the slope and shape of these functions. Through the application of these discrete transformation functions, the uniform input domain can be mapped to the discretized domain.

4.1. Modified Arctangent Function

The arctangent function can discretely map the original numerical values to a certain interval. The range of the arctangent function a r c t a n ( x ) is between π / 2 and π / 2 , while the slope of this function at the zero point is 1. Hence, we need to scale the range of the original arctangent function and shift the center of the function to meet the mapping range requirements and set the slope at (0.5, 0.5). The modified arctangent function is as follows:
mod - a r c t a n ( x ) = 1 π a r c t a n ( 0.25 π s c a l e ( x 0.5 ) ) + 0.5
The modified arctangent function can suppress the input vectors when they are smaller than 0.5 and activate them when they are larger than 0.5. The value of s c a l e can be set to adjust the slope of the modified arctangent function and obtain different discrete transformation effects; a larger value of s c a l e results in a more pronounced discrete transformation. However, the range of output values is smaller than one and the effects of activation and inhibition are not prominent. As a result, it is necessary to modify the logistic function, which is commonly used in neural networks and has a wide output range and remarkable activation effects.

4.2. Modified Logistic Function

The common logistic function is defined as follows:
σ ( x ) = 1 1 + exp ( x )
This function can be considered an extrusion function that extrudes the input from a real field to a range between zero and one. When the input value is near zero, the logistic function is approximately linear. Input values are suppressed when they are sufficiently small. The smaller the input, the closer it is to zero; the larger the input, the closer it is to one. This behavior somewhat resembles that of biological neurons, which respond to some inputs with an output of one and suppress other inputs with an output of zero. Compared with the step activation function used by the perceptron, the logistic function is continuous and has better mathematical properties.
Due to the nature of the logistic function, the neuron equipped with the logistic activation function has the following two properties: first, its expression can be directly regarded as a probability distribution function, so that the neural network can be better combined with the statistical learning model; second, it can be regarded as a soft gate that controls the amount of information output from other neurons. To ensure that the input and output range is between zero and one and the slope of the function is adjustable at 0.5, we modify the logistic function as follows:
mod - σ ( x ) = 1 1 + exp ( s c a l e ( x 0.5 ) )
where s c a l e is the slope adjustment parameter used to adjust the slope of the function at zero. The large value of s c a l e represents the very large slope of the function at 0.5.

4.3. Modified Hard-Logistic Function

The logistic and arctangent functions are both "s"-shaped functions with saturation characteristics, but the calculation cost is high, as these two functions are nonlinear in the saturation range at the left and right ends. Therefore, the logistic function can be approximated by piecewise functions. The derivative of the logistic function σ ( x ) is σ ( x ) = σ ( x ) ( 1 σ ( x ) ) . The first-order Taylor expansion near zero of σ ( x ) is g l ( x ) , which is approximated by σ ( 0 ) + x × σ ( 0 ) = 0.25 x + 0.5 . It, therefore, follows that the function can be approximated by a piecewise hard-logistic function, which is defined by the equation below.
hard - logistic ( x ) = 1 , x 2 g l ( x ) , 2 < x < 2 0 , x 2 = max min g l ( x ) , 1 , 0 = max ( min ( 0.25 x + 0.5 , 1 ) , 0 )
The s c a l e parameter also needs to be introduced into the hard-logistic function. To guarantee that the slope of the function is adjustable at 0.5 and the input and output range is between zero and one, we modify the hard-logistic function as follows:
mod - hard - logistic ( x ) = 1 , x x 2 g l ( s c a l e ( x 0.5 ) ) , x 1 < x < x 2 0 , x x 1 = max min g l ( s c a l e ( x 0.5 ) ) , 1 , 0 = max ( min ( 0.25 s c a l e ( x 0.5 ) + 0.5 , 1 ) , 0 )
where x 1 is 2 / s c a l e + 0.5 and x 2 is 2 / s c a l e + 0.5 . Here, x 1 is the left turning point and x 2 is the right turning point of the mod-hard-logistic function.
With these discrete transformation functions, the unified input domain can be mapped to the binarized domain between zero and one. The slope at 0.5 of the mod-arctangent function is the same as that of mod- σ ( x ) and mod-hard-logistic(x). These three discrete transformation functions all have activation and inhibition effects. The function graphs of these functions with different s c a l e values are in plotted Figure 2.

5. Simulation and Results

In this section, simulation and experimental results are presented to demonstrate the practicality and effectiveness of the proposed model on a deep learning platform with Python coding. The symbol rate and sampling rate of generated signals are considered, respectively. These signals exhibit some deviations, such as sampling rate deviation, carrier frequency offset, and the effects of AWGN. The modulation types of the generated signals include 4ASK, 8ASK, 2PSK, 4PSK, 8PSK, 16QAM, and 64QAM. These signals can simulate real-world signals with the relevant parameters. The number of samples for clustering is 2000 for each modulation type, such that the total number of samples is 14,000. The SNR value of the environment ranges from −5 dB to 10 dB. The signals are also affected by additive noise. The amplitude value of this kind of noise obeys a Gaussian distribution, while the power spectral density obeys a uniform distribution. The multiple interference factors convey the difficulty of distinguishing the specific modulation types. The different cumulants and amplitude moments in the constructed feature space can help to overcome these interference factors.

5.1. Clustering Results

We first use the root layer to roughly cluster the 14,000 input samples. The length and width of this layer are both 50. From the description in Section 2, we know that the | C 42 | features of MASK, MPSK, and MQAM are different; thus, | C 42 | can be the classification feature of the root layer. The root layer maps the | C 42 | of the input samples and clusters them in different regions. The learning rate of modified activation functions η ( 0 ) is set to one. The clustering images in the root layer for the seven kinds of samples when the SNR is 10 dB are presented in Figure 3a. It is evident that the boundaries of 2PSK, 4ASK, and 8ASK are clear while the boundaries of 4PSK, 8PSK, 16QAM, and 64QAM are fuzzy. We, therefore, mark the latter group with four lags, namely 2PSK, MASK (including 4ASK and 8ASK), MPSK (including 4PSK and 8PSK), and MQAM (including 16QAM and 64QAM) in Figure 3b.
The classification accuracy and the quantization error of the MASK, MPSK, and MQAM feature samples clustered in the root layer are presented in Figure 4. Here, classification accuracy is defined as the ratio between the number of correctly classified samples and the total number of feature samples. The quantization error is the average distance between each input sample and its best matching unit. As the number of iterations increases, the classification accuracy of the root layer climbs up before 7000 iterations, then fluctuates between 7000 and 18,000 iterations. The classification accuracy reaches its maximum of 97.6% when the number of iterations is 14,000. Moreover, the quantization error climbs from 0.542 to 0.832 when the number of iterations ranges from 1000 to 2000, then drops from 0.383 to 0.016 as the iterations increase from 3000 to 12,000. When the number of iterations exceeds 12,000, the quantization error converges to a fairly low value, less than the parameter τ = 0.05 , and remains at this level. At this range, the classification accuracy also converges to between 95.6% and 97.6% and does not increase further.
To ensure that the MASK, MPSK, and MQAM samples cluster with clear boundaries, the hierarchical SOM network grows new leaf layers based on the root layer to distinguish between the candidate samples with different modulation orders. The leaf layer for MPSK clusters the samples with the second feature | C 40 | , while that for MQAM clusters the samples with the third feature μ 42 . The output of the leaf layer is shown in Figure 5, including two categories of candidate MPSK (4PSK and 8PSK) and MQAM (16QAM and 64QAM) data; here, the length of the network is also 50. The clustering results with a clear boundary for MPSK samples in the leaf layers are presented in Figure 5a. The original clustering results for MQAM samples without modified activation functions in the leaf layers are presented in Figure 5b.
However, the 16QAM and 64QAM samples cluster closely, which introduces bad clustering effects. Consequently, we apply a discrete transformation with modified activation functions to make the input vectors discretize and cluster more explicitly. The numerical size relationship between two kinds of QAM samples does not change due to the monotonicity of discrete transformation. Figure 6 illustrates the classification accuracy of the leaf layers without discrete transformation. The accuracy for 4PSK and 8PSK reaches 99.5% when the number of iterations is 1000 and remains high as iterations increase. Moreover, the accuracy for 16QAM and 64QAM is 94.1% at the beginning of iterations and increases to its maximum value of 97.9% when the number of iterations reaches 16,000.
The clustering results of QAM samples with different orders through various modified activation functions with two distinct values of s c a l e are presented in Figure 7. From the six contrasting images, we can observe that as the value of s c a l e increases, the boundaries between two clusters become clear and the winning neurons in different regions tend to be more discretized. The clustering results of the mod-arctangent function are worse than those of the mod-logistic function, while the mod-hard-logistic function can be seen to have the best clustering effect, as the edges of the two clusters are far away from each other. This phenomenon results from the binarization characteristics of different activation functions near the minimum and maximum points. Most of the output data is clustered in such a way as to make the clustering results obvious. The classification accuracy of these three activation functions for discrete transformation is shown in Figure 8 when s c a l e is 40. The accuracies with different numbers of iterations all exceed 98%. In general, the accuracy of clustering results obtained using the mod-arctangent function is lower than that obtained through the mod-logistic function, while the mod-hard-logistic function has the highest accuracy.
The experimental results above are under the SNR condition of 10 dB. As the SNR value ranges from −5 dB to 10 dB, the classification accuracies in the root layer present a trend from low value to high value, as shown in Figure 9. When the SNR is lower than −1 dB, the classification accuracies are smaller than 70%. When the SNR is between −1 dB and 1 dB, the classification accuracies increase unsteadily to 94.6%. Subsequently, the classification accuracies rise smoothly from 93.8% to 97.3% when the SNR ranges from 2 dB to 10 dB. As for the leaf layer results in Figure 10, the classification accuracies for MPSK samples remain high from −5 dB to 10 dB. The classification accuracies for MQAM samples without discrete transformation increase from 88.3% to 97.3% when the SNR ranges between −5 dB and −1 dB, then fluctuate between 96.2% and 99.6% when the SNR increases towards 10 dB. The classification accuracies for MQAM samples with three discrete transformations increase from −5 dB to 1 dB, then fluctuate between 2 dB and 10 dB, as shown in Figure 11. The accuracies obtained with the mod-hard-logistic function usually exceed those obtained using the mod-logistic function, followed by those obtained with the mod-arctangent function.
Compared with previous research into automatic modulation classification on four modulation categories, including 2PSK, 4PSK, 8PSK, and 16QAM, and using the methods of one sample 2D K-S classifier with parameter estimated by ECM (K-S classifier), a combination of genetic programming and k-nearest neighbor (GP-KNN classifier), clustering analysis of constellation signature (CACS classifier) [39], and maximum likelihood classifier with parameters estimated by SQUAREM-PC (ML classifier), our proposed hierarchical SOM model (H-SOM) can deal with more modulation types and obtain higher accuracy under the same SNR environment, as shown in Table 2. The methods in prior works extract spectral and statistical features, which can enhance the differences between the samples of interest. In our method, however, the hierarchical SOM model selects only statistical features, including high-order cumulants C 42 and amplitude moments μ 42 , based on recent research. C 42 can suppress the side-effects of additive Gaussian noise and demonstrates robustness when classifying MASK and MPSK signals. μ 42 reflects the level of compactness for the received signals and works as an indicator to facilitate the classification of 16QAM and 64QAM signals. H-SOM is sufficiently sensitive to present the differences in input data in the output clustering layer.

5.2. Comparison of Computational Requirements

In practical applications, computational requirements play a crucial role when applying the proposed model for the inference stage after training. The inference stage is based on the existing parameters in the training stage. When training a sufficiently large amount of data to obtain a trustworthy model, it is necessary to consider the consumption of computational resources in order to reduce time costs and accelerate the inference stage. For some time-first scenarios, such as embedded applications and quick response systems, it is favorable to select the models with lower computational requirements to obtain better performance and reduce power consumption. When calculating the computational requirements, the mathematical operations applied to complete the classification are taken as the measure of the computational requirements for each classifier. We further make some assumptions to calculate the number of various operations as follows: (1) the total number of modulation types for classification is T m ; (2) the number of samples in the i-th modulation candidate is M i ; (3) the signal sample length is N; (4) the number of weights in the model is W; (5) the number of proposed neurons in the i-th layer of hierarchical SOM is P i . The overall computational requirements of the maximum likelihood classifier (ML) and stacked auto-encoder-based deep neural network (SAE-DNN) classifiers are presented in [40], while the computational requirements of the inference stage of hierarchical SOM are simply related to the addition operations in the competition procedure, which are presented in Table 3. There are no extra computational resources required for the samples of 2PSK and MASK in the leaf layer.
We can observe that classifiers based on the ML theorem and SAE-DNN require multiplications and exponential operations. As the resource consumption of addition operations is far lower than that for multiplications and exponential operations and the number of proposed neurons is quite smaller than the models for comparison, we can make a conclusion that the computational requirements, including additions, multiplications, exponents, and logarithms, of the hierarchical SOM model are smaller than those of the other two models.

6. Discussion

The hierarchical SOM model can present the differences between groups of data and cluster similar features on a two-dimensional plane. The main advantage of using a hierarchical SOM is the clear visualization of clustering results. The mapping of feature samples in grid form makes it easy to present the similarities and clustering results of the data. SOM processes all data in the feature space to generate independent clusters. The nonlinear classification ability of SOM makes samples cluster in certain regions through the application of self-learning procedures. SOM can handle several classification problems by providing a functional, interactive, and intelligible outline of the data. The hierarchical SOM can cluster distinct samples roughly in the root layer, then cluster samples with subtle differences in the leaf layer. This model can cluster MQAM and MPSK samples based on the HOC and amplitude moment features of the input digital signals. The hierarchical SOM has the advantages of higher classification accuracy and lower computational resource consumption. Moreover, the clustering model is explainable, which makes the clustering results trustworthy according to the theoretical values of selected features. The samples with low feature values gather in the lower left corner of the output layer, while those with high values gather in the upper right corner of the output layer. The boundary line of clustering usually runs from the upper left to the lower right. We can find that the sums of horizontal and vertical coordinates are positively correlated with the feature values of samples.
One disadvantage of a hierarchical SOM is that it requires a sufficient number of iterations to generate clusters and obtain high accuracy. The weight vectors need to iterate more than 10,000 times to distinguish samples with high accuracy. The lack of adequate iterations for weight vectors will result in fuzzy boundaries of clusters and fluctuating classification accuracy. Moreover, it is also uncertain which iteration produces the highest accuracy, making it necessary to try enough iterations to choose the most eligible one in practical applications. Another disadvantage of hierarchical SOM is that the clustering results depend on the value distribution of the original dataset. Similar samples belonging to different categories may cause some samples to cluster in adjacent regions or even mix up in the same region. This phenomenon will unavoidably decrease the classification accuracy of the hierarchical SOM model. While the discrete transformation proposed in this paper can reduce the negative impacts of similar samples in different categories, it cannot eliminate this phenomenon completely. It is, therefore, practical to remove abnormal samples before using the hierarchical SOM model to get better clustering results in the output layer before the normalization of the features.

7. Conclusions

The hierarchical SOM model proposed in this paper applies stacked neuron layers to cluster MASK, MPSK, and MQAM digital samples based on the normalized HOC and amplitude moment features without the need to set thresholds based on these features. These features are robust to noise and show the differences between different modulation types. Unlike conventional one-layer SOM models, the hierarchical SOM model can generate leaf layers in the procedure of layering according to the values of quantization errors. This model can roughly cluster the MASK, MPSK, and MQAM samples in the root layer and then finely cluster the MPSK and MQAM samples in the leaf layers. In addition, this model can improve the visualization and classification results of clustering MQAM samples based on discrete transformation in the leaf layer. The classification accuracies of hierarchical SOM are higher than those of the four existing clustering methods selected for comparison. While hierarchical SOM is suitable for fine classification of some similar data, the model training parameters need much professional experience; this model requires a sufficient number of iterations in order for the clustering procedures to obtain high accuracy. In future work, it would be feasible to consider the application of this hierarchical SOM model in different clustering problems focusing on more categories of modulated signals or other types of similar data.

Author Contributions

Conceptualization, Z.L.; funding acquisition, Q.W. and Z.X.; investigation, Z.L.; methodology, Z.L. and Q.W.; resources, Y.Z. and Q.W.; software, Q.W.; supervision, Y.Z.; writing—original draft, Z.L.; writing—review and editing, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 62002365 and 61874140.

Acknowledgments

We appreciate our reviewers and editors for their precious time.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, X.Y.; Yang, D.Y.; Gamal, A.E. Deep neural network architectures for modulation classification. In Proceedings of the 2017 IEEE 51st Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 29 October–1 November 2017; pp. 915–919. [Google Scholar] [CrossRef]
  2. Huang, L.; Zhang, Y.; Pan, W.; Chen, J.; Qian, L.P.; Wu, Y. Visualizing Deep Learning-based Radio Modulation Classifier. IEEE Trans. Cogn. Commun. Netw. 2021, 7, 47–58. [Google Scholar] [CrossRef]
  3. Dobre, O. Signal identification for emerging intelligent radios: Classical problems and new challenges. IEEE Instrum. Meas. Mag. 2015, 18, 11–18. [Google Scholar] [CrossRef]
  4. Dulek, B.; Ozdemir, O.; Varshney, P.K.; Su, W. Distributed Maximum Likelihood Classification of Linear Modulations Over Nonidentical Flat Block-Fading Gaussian Channels. IEEE Trans. Wirel. Commun. 2015, 14, 724–737. [Google Scholar] [CrossRef]
  5. Hazza, A.; Shoaib, M.; Alshebeili, S.A.; Fahad, A. An overview of feature-based methods for digital modulation classification. In Proceedings of the 2013 IEEE 1st International Conference on Communications, Signal Processing and Their Applications (ICCSPA), Sharjah, United Arab Emirates, 12–14 February 2013; pp. 1–6. [Google Scholar] [CrossRef]
  6. Li, X.F.; Dong, F.W.; Zhang, S.; Guo, W.B. A Survey on Deep Learning Techniques in Wireless Signal Recognition. Wirel. Commun. Mob. Comput. 2019, 2019, 5629572. [Google Scholar] [CrossRef]
  7. Bari, M.; Khawar, A.; Doroslovacki, M.; Clancy, T.C. Recognizing FM, BPSK and 16-QAM using supervised and unsupervised learning techniques. In Proceedings of the 2015 49th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 8–11 November 2015; pp. 160–163. [Google Scholar] [CrossRef]
  8. Liu, W.; Li, Y. Constellation based signal modulation recognition for MQAM. In Proceedings of the 2017 IEEE 9th International Conference on Communication Software and Networks (ICCSN), Guangzhou, China, 6–8 May 2017; pp. 826–829. [Google Scholar] [CrossRef]
  9. Huang, S.; Dai, R.; Huang, J.; Yao, Y.; Gao, Y.; Ning, F.; Feng, Z. Automatic Modulation Classification Using Gated Recurrent Residual Network. IEEE Internet Things J. 2020, 7, 7795–7807. [Google Scholar] [CrossRef]
  10. Peng, S.; Sun, S.; Yao, Y.D. A Survey of Modulation Classification Using Deep Learning: Signal Representation and Data Preprocessing. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–9. [Google Scholar] [CrossRef] [PubMed]
  11. Yan, X.; Zhang, G.; Wu, H.C.; Liu, G. Automatic modulation classification in α-stable noise using graph-based generalized second-order cyclic spectrum analysis. Phys. Commun. 2019, 37, 100854. [Google Scholar] [CrossRef]
  12. Dong, S.; Li, Z.; Zhao, L. A Modulation Recognition Algorithm based on Cyclic Spectrum and SVM Classification. In Proceedings of the 2020 IEEE 4th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chongqing, China, 12–14 June 2020; pp. 2123–2127. [Google Scholar] [CrossRef]
  13. Dobre, O.A.; Oner, M.; Rajan, S.; Inkol, R. Cyclostationarity-Based Robust Algorithms for QAM Signal Identification. IEEE Commun. Lett. 2012, 16, 12–15. [Google Scholar] [CrossRef]
  14. Zhang, L.; Li, Q. Research on Modulation Recognition Algorithm Based on Higher-Order Cumulant. J. Inf. Eng. Univ. 2017, 18, 403–408. [Google Scholar]
  15. Al-Nuaimi, D.H.; Hashim, I.A.; Abidin, I.; Salman, L.B.; Isa, N. Performance of Feature-Based Techniques for Automatic Digital Modulation Recognition and Classification: A Review. Electronics 2019, 8, 1407. [Google Scholar] [CrossRef]
  16. Li, Y.L.; Li, B.B.; Yin, C.Y. Modulation classification of MQAM signals using particle swarm optimization and subtractive clustering. In Proceedings of the 2010 IEEE International Conference on Signal Processing, Beijing, China, 24–28 October 2010; pp. 1537–1540. [Google Scholar] [CrossRef]
  17. Kaur, M.; Ahmed, P.; Soni, A.K.; Rafiq, M.Q. Wavelet Transform use for P300 Signal Clustering by Self-Organizing Map. Int. J. Sci. Eng. Res. 2013, 4, 1631–1638. [Google Scholar]
  18. Zhou, L.; Cai, Q.; He, F.M.; Man, H. MSOM based automatic modulation recognition and demodulation. In Proceedings of the 34th IEEE Sarnoff Symposium, Princeton, NJ, USA, 3–4 May 2011; pp. 1–5. [Google Scholar] [CrossRef]
  19. Xu, Y.Q.; Ge, L.D.; Wang, B. Digital Modulation Recognition Method Based on Self-Organizing Map Neural Networks. In Proceedings of the 2008 4th International Conference on Wireless Communications, Networking and Mobile Computing, Dalian, China, 12–14 October 2008; pp. 1–4. [Google Scholar] [CrossRef]
  20. Kiang, M.Y. Extending the Kohonen self-organizing map networks for clustering analysis. Comput. Stat. Data Anal. 2002, 38, 161–180. [Google Scholar] [CrossRef]
  21. Chaudhary, V.; Ahlawat, A.K.; Bhatia, R.S. An efficient Self-organizing map learning algorithm with winning frequency of neurons for clustering application. In Proceedings of the 2013 IEEE 3rd International Advance Computing Conference (IACC), Ghaziabad, India, 22–23 February 2013; pp. 672–676. [Google Scholar] [CrossRef]
  22. Kohonen, T. Essentials of the self-organizing map. Neural Netw. Off. J. Int. Neural Netw. Soc. 2013, 37, 52–65. [Google Scholar] [CrossRef]
  23. Melin, P.; Monica, J.C.; Sanchez, D.; Castillo, O. Analysis of Spatial Spread Relationships of Coronavirus (COVID-19) Pandemic in the World using Self Organizing Maps. Chaos Solitons Fractals 2020, 138, 109917–109923. [Google Scholar] [CrossRef]
  24. Braga, P.H.M.; Bassani, H.F. A Semi-Supervised Self-Organizing Map for Clustering and Classification. In Proceedings of the 2018 IEEE International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar] [CrossRef]
  25. Garay, A.B.; Contreras, G.P.; Patio-Escarcina, R.E. A GH-SOM optimization with SOM labelling and dunn index. In Proceedings of the 11th International Conference on Hybrid Intelligent Systems, Melacca, Malaysia, 5–8 December 2011; pp. 572–577. [Google Scholar] [CrossRef]
  26. Huang, S.Y.; Huang, Y. Network Traffic Anomaly Detection Based on Growing Hierarchical SOM. In Proceedings of the 43rd Annual IEEE/IFIP International Conference on Dependable Systems and Networks, Hungary, Budapest, 24–27 June 2013; pp. 1–2. [Google Scholar] [CrossRef]
  27. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over the Air Deep Learning Based Radio Signal Classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef]
  28. Ali, A.; Yangyu, F. Unsupervised feature learning and automatic modulation classification using deep learning model. Phys. Commun. 2017, 25, 75–84. [Google Scholar] [CrossRef]
  29. Swami, A.; Sadler, B.M. Hierarchical digital modulation classification using cumulants. IEEE Trans. Commun. 2000, 48, 416–429. [Google Scholar] [CrossRef]
  30. Li, C.; Xiao, J.; Xu, Q. A novel modulation classification for PSK and QAM signals in wireless communication. In Proceedings of the IET International Conference on Communication Technology and Application (ICCTA 2011), Beijing, China, 14–16 October 2011; pp. 89–92. [Google Scholar] [CrossRef]
  31. Han, Y.; Wei, G.; Song, C.; Lai, L. Hierarchical digital modulation recognition based on higher-order cumulants. In Proceedings of the 2012 Second International Conference on Instrumentation, Measurement, Computer, Communication and Control, Harbin, China, 8–10 December 2012; pp. 1645–1648. [Google Scholar] [CrossRef]
  32. Zhang, H.; Lou, H. Automatic modulation recognition algorithm for MQAM signal. J. Commun. 2019, 40, 200–211. [Google Scholar] [CrossRef]
  33. Kohonen, T. Self-Organizing Maps, 3rd ed.; Springer Series in Information Sciences; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar] [CrossRef]
  34. Taşdemir, K.; Milenov, P.; Tapsall, B. Topology-Based Hierarchical Clustering of Self-Organizing Maps. IEEE Trans. Neural Netw. 2011, 22, 474–485. [Google Scholar] [CrossRef]
  35. Cheng, H.; Han, H.; Wu, L.; Chen, L. A 1-Dimension Structure Adaptive Self-Organizing Neural Network for QAM Signal Classification. In Proceedings of the Third International Conference on Natural Computation (ICNC 2007), Haikou, China, 24–27 August 2007; Volume 3, pp. 53–57. [Google Scholar] [CrossRef]
  36. Fu, J.; Che, G. Fusion Fault Diagnosis Model for Six-Rotor UAVs Based on Conformal Fourier Transform and Improved Self-Organizing Feature Map. IEEE Access 2021, 9, 14422–14436. [Google Scholar] [CrossRef]
  37. Riese, F.M.; Keller, S.; Hinz, S. Supervised and Semi-Supervised Self-Organizing Maps for Regression and Classification Focusing on Hyperspectral Data. Remote Sens. 2020, 12, 7. [Google Scholar] [CrossRef]
  38. Musci, M.; Parigi, G.; Cantoni, V.; Piastra, M. A scalable multi-signal approach for the parallelization of self-organizing neural networks. Neural Netw. Off. J. Int. Neural Netw. Soc. 2020, 123, 108–117. [Google Scholar] [CrossRef]
  39. Jajoo, G.; Kumar, Y.; Yadav, S.K. Blind Signal PSK/QAM Recognition Using Clustering Analysis of Constellation Signature in Flat Fading Channel. IEEE Commun. Lett. 2019, 23, 1853–1856. [Google Scholar] [CrossRef]
  40. Shah, M.H.; Dang, X. A Robust Approach for AMC in Frequency Selective Fading Scenarios using Unsupervised Sparse Autoencoder based Deep Neural Network. IET Commun. 2019, 13, 423–432. [Google Scholar] [CrossRef]
Figure 1. The structure of the hierarchical SOM model.
Figure 1. The structure of the hierarchical SOM model.
Sensors 22 06449 g001
Figure 2. The modified functions with different values of scale.
Figure 2. The modified functions with different values of scale.
Sensors 22 06449 g002
Figure 3. Clustering results in the root layer with different labels.
Figure 3. Clustering results in the root layer with different labels.
Sensors 22 06449 g003
Figure 4. Classification accuracy and quantization error of the root layer.
Figure 4. Classification accuracy and quantization error of the root layer.
Sensors 22 06449 g004
Figure 5. Clustering results of the leaf layers for MPSK and MQAM without discrete transformation.
Figure 5. Clustering results of the leaf layers for MPSK and MQAM without discrete transformation.
Sensors 22 06449 g005
Figure 6. Classification accuracy of the leaf layers for MPSK and MQAM without discrete transformation.
Figure 6. Classification accuracy of the leaf layers for MPSK and MQAM without discrete transformation.
Sensors 22 06449 g006
Figure 7. Clustering results of the leaf layer for MQAM with discrete transformation and different scales.
Figure 7. Clustering results of the leaf layer for MQAM with discrete transformation and different scales.
Sensors 22 06449 g007
Figure 8. Classification accuracy of the leaf layer for MQAM with discrete transformation.
Figure 8. Classification accuracy of the leaf layer for MQAM with discrete transformation.
Sensors 22 06449 g008
Figure 9. Classification accuracy and quantization error of the root layer under different SNRs.
Figure 9. Classification accuracy and quantization error of the root layer under different SNRs.
Sensors 22 06449 g009
Figure 10. Classification accuracy of the leaf layers for MPSK and MQAM without discrete transformation under different SNRs.
Figure 10. Classification accuracy of the leaf layers for MPSK and MQAM without discrete transformation under different SNRs.
Sensors 22 06449 g010
Figure 11. Classification accuracy of the leaf layer for MQAM with discrete transformation under different SNRs.
Figure 11. Classification accuracy of the leaf layer for MQAM with discrete transformation under different SNRs.
Sensors 22 06449 g011
Table 1. Index range and theoretical values of features of different modulation types.
Table 1. Index range and theoretical values of features of different modulation types.
TypesIndex Range | C 42 | | C 40 | μ 42
4ASK[1, S]1.36E 2 --
8ASK[S+1, 2S]1.24E 2 --
2PSK[2S+1, 3S]2E 2 --
4PSK[3S+1, 4S]E 2 E 2 -
8PSK[4S+1, 5S]E 2 0-
16QAM[5S+1, 6S]0.68E 2 -2.3590
64QAM[6S+1, 7S]0.62E 2 -2.5025
Table 2. The accuracy results among different models.
Table 2. The accuracy results among different models.
Model NameModulation TypesSNR/dBAccuracy/%
K-S classifier2PSK, 4PSK, 8PSK, 16QAM1069.8
GP-KNN classifier2PSK, 4PSK, 8PSK, 16QAM1070.9
CACS classifier2PSK, 4PSK, 8PSK, 16QAM1085.2
ML classifier2PSK, 4PSK, 8PSK, 16QAM1089.3
H-SOM classifierMASK, MPSK, MQAM (7 types)1096.6
Table 3. Computational requirements of different classifiers.
Table 3. Computational requirements of different classifiers.
ClassifiersAdditionsMultiplicationsExponentsLogarithms
ML 6 N T m i = 1 T m M i 5 N T m i = 1 T m M i N T m i = 1 T m M i N T m
SAE-DNN N W T m i = 1 T m M i N W T m i = 1 T m M i N T m i = 1 T m M i 0
H-SOM P 1 i = 1 T m M i + P 2 i = 2 T m M i 000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Z.; Wang, Q.; Zhu, Y.; Xing, Z. Automatic Modulation Classification for MASK, MPSK, and MQAM Signals Based on Hierarchical Self-Organizing Map. Sensors 2022, 22, 6449. https://doi.org/10.3390/s22176449

AMA Style

Li Z, Wang Q, Zhu Y, Xing Z. Automatic Modulation Classification for MASK, MPSK, and MQAM Signals Based on Hierarchical Self-Organizing Map. Sensors. 2022; 22(17):6449. https://doi.org/10.3390/s22176449

Chicago/Turabian Style

Li, Zerun, Qinglin Wang, Yufei Zhu, and Zuocheng Xing. 2022. "Automatic Modulation Classification for MASK, MPSK, and MQAM Signals Based on Hierarchical Self-Organizing Map" Sensors 22, no. 17: 6449. https://doi.org/10.3390/s22176449

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop