Next Article in Journal
Analysis of Train–Track–Bridge Coupling Vibration Characteristics for Heavy-Haul Railway Based on Virtual Work Principle
Previous Article in Journal
Development of Photonic Multi-Sensing Systems Based on Molecular Gates Biorecognition and Plasmonic Sensors: The PHOTONGATE Project
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on None-Line-of-Sight/Line-of-Sight Identification Method Based on Convolutional Neural Network-Channel Attention Module

1
State Key Laboratory of Satellite Navigation System and Equipment Technology, Shijiazhuang 050081, China
2
The 54th Research Institute of China Electronics Technology Group Corporation, Shijiazhuang 050081, China
3
School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(20), 8552; https://doi.org/10.3390/s23208552
Submission received: 24 August 2023 / Revised: 11 October 2023 / Accepted: 13 October 2023 / Published: 18 October 2023

Abstract

:
None-Line-of-Sight (NLOS) propagation of Ultra-Wideband (UWB) signals leads to a decrease in the reliability of positioning accuracy. Therefore, it is essential to identify the channel environment prior to localization to preserve the high-accuracy Line-of-Sight (LOS) ranging results and correct or reject the NLOS ranging results with positive bias. Aiming at the problem of the low accuracy and poor generalization ability of NLOS/LOS identification methods based on Channel Impulse Response (CIR) at present, the multilayer Convolutional Neural Networks (CNN) combined with Channel Attention Module (CAM) for NLOS/LOS identification method is proposed. Firstly, the CAM is embedded in the multilayer CNN to extract the time-domain data features of the original CIR. Then, the global average pooling layer is used to replace the fully connected layer for feature integration and classification output. In addition, the public dataset from the European Horizon 2020 Programme project eWINE is used to perform comparative experiments with different structural models and different identification methods. The results show that the proposed CNN-CAM model has a LOS recall of 92.29%, NLOS recall of 87.71%, accuracy of 90.00%, and F1-score of 90.22%. Compared with the current relatively advanced technology, it has better performance advantages.

1. Introduction

With the rapid development of Internet of Things (IoT) technology, intelligent mobile terminal technology, and mobile computing technology, the demand for indoor location services in many industries is getting higher, and the demand for real-time positioning of personnel is becoming more urgent [1,2]. Ultra-wideband (UWB) technology stands out among wireless positioning technologies because of its low power consumption, high-ranging accuracy, high temporal resolution, strong anti-interference capability, etc. [3]. Nevertheless, it is limited by multi-user interference, clock drift, frequency drift, and Non-Line-of-Sight (NLOS) propagation in practical application scenarios [4]. Signal propagation in the NLOS state is affected by obstacles that increase the arrival time, thus causing a positive bias in the distance measurement. It is considered to be one of the main challenges faced by high-precision positioning systems. Therefore, NLOS and Line-of-Sight (LOS) identification before positioning is critical [5].
NLOS/LOS identification methods can be divided into three categories: distance-based estimation methods, position-based estimation methods, and CIR-based estimation methods. The distance-based estimation method performs NLOS and LOS identification by detecting the variance of multiple distance measurements at a given location or by detecting whether the current distance measurement conforms to a specified distribution [6]. The method is relatively simple, but it is limited by the distribution function or time delay. The position-based estimation method performs NLOS/LOS identification during the position-solving process or after the position computation is completed. NLOS can be identified by comparing the position estimates generated using different subsets of the distance estimates when redundant ranging information is present. However, this method is ineffective in the absence of redundant ranging information [7]. If additional environmental information, such as maps, geometric relationships, and path continuity, existed, NLOS identification can be performed using position constraint analysis after the position calculation is completed. However, the disadvantages of external information sources can limit the system’s stability and increase complexity. Over time, CIR-based estimation methods have received extensive attention from scholars [8,9,10]. CIR reflects the fluctuation and fading of the signal in the channel environment, so the channel parameters of CIR are combined with joint likelihood function [11], machine learning [12], threshold comparison [13], and deep learning [14,15,16] to identify NLOS/LOS.
In this paper, a novel method for NLOS/LOS identification based on CIR combined with deep learning is proposed. This method aims to improve NLOS/LOS identification accuracy and reduce computational complexity. Our main contributions are summarized as follows:
  • A multilayer Convolutional Neural Network (CNN) combined with a Channel Attention Module (CAM) for the NLOS/LOS identification method is proposed. The method takes the One-dimensional CIR signal as input, uses three groups of convolution modules (Convolution + BN + ReLU + Max-pooling) and CAM for self-extraction of key features, and the global average pooling layer is used to replace the fully connected layer for feature integration and classification output, which achieves NLOS/LOS identification.
  • Two schemes are proposed on how to determine the specific structure of the CNN-CAM network and how to determine the optimal parameters. In the first scheme, the proposed CNN-CAM model is compared with CNN and CNN-CAM models with different structures, and it aims to select the optimal model structure for NLOS/LOS identification. In the second scheme, the effect of different learning rates and batches on the identification accuracy is compared experimentally for the proposed model, and it aims to determine the optimal parameters of the model.
  • A scheme on how to verify the superiority of the proposed CNN-CAM method is offered. Firstly, the public dataset of the European Horizon 2020 program project eWINE is visualized and analyzed to illustrate the feasibility of using this dataset for experiments. Then, comparative experiments of several machine learning and deep learning identification methods are conducted using the dataset to validate the state-of-the-art of the proposed CNN-CAM method.
The rest of the paper is organized as follows. Section 2 introduces the related work of other scholars in the field of CIR-based NLOS/LOS identification. Section 3 analyzes the key problems of NLOS/LOS and the performance of CIR in the NLOS/LOS environment. They provide new ideas for the design of the proposed model. Section 4 describes the details of the proposed NLOS/LOS identification method based on CNN-CAM. Section 5 performs a series of visualizations on the CIR dataset and designs various experiments to evaluate the performance of the proposed method. Section 6 summarizes the article and discusses the advantages and limitations of this approach and future work.

2. Related Work

As mentioned in Section 1, NLOS/LOS identification methods are classified into three categories. Among them, CIR-based NLOS/LOS identification methods have received extensive attention from scholars. In this paper, we focus on this class of methods.
The channel parameters proposed in the existing literature mainly include kurtosis, skewness, maximum amplitude, peak time, rise time, total energy, mean excess delay, root-mean-square (RMS) delay spread, saturation, peak-to-average ratio, received signal power, energy steep rise amplitude, false crests number, first path error, first path distance error, etc. [9,10]. Ref. [11] proposed a method to identify NLOS and LOS using three-channel parameters, kurtosis, RMS delay spread, and mean excess delay, as statistical information and by building a likelihood function. Marano et al. extracted six channel parameters from CIR waveforms, including kurtosis, received signal power, maximum amplitude, rise time, mean excess delay, and RMS delay spread, then used Least Squares Support Vector Machine (LS-SVM) for NLOS and LOS identification [12]. However, the method does not consider the correlation between channel parameters. Li et al. proposed a new method that takes the sum of peak time and rise time as a new channel parameter and combines it with the number of undetected peaks [13]. The method has a high identification accuracy when the threshold is selected appropriately, but it is prone to identification errors when the peak time and rise time differ significantly from the expected or when the threshold is selected incorrectly. NLOS/LOS identification is essentially a binary classification problem. Additionally, the above methods need to manually extract the channel parameters of the CIR for NLOS/LOS identification, which can lead to a system whose reliability and robustness cannot be guaranteed.
In recent years, with the rapid development of deep neural networks, NLOS/LOS identification methods based on deep learning have received extensive attention from scholars. This method uses CIR data as system input and accomplishes feature extraction employing model self-learning. Jiang et al. used invertible transform for denoising the CIR dataset (European Horizon 2020 Programme project eWINE) and used CNN models to identify NLOS. The identification accuracy of the method is up to 81.68% [14]. Jiang et al. trained and tested the Convolutional Neural Network and Long Short-Term Memory (CNN-LSTM) model using the CIR dataset (European Horizon 2020 Programme project eWINE) as input. The method achieves an identification rate of up to 82.14% [15]. Li et al. proposed a method that takes the real and imaginary parts of the original CIR and its Fourier transform as inputs and utilizes a three-channel Convolutional Neural Network and Bidirectional Long Short-Term Memory (CNN-BiLSTM) for identification. The article used public datasets (European Horizon 2020 Programme project eWINE) to verify and found that the proposed method has an identification accuracy of 85.71% and outperforms both LSTM and CNN-LSTM [16]. Pei et al. proposed a Fully Convolution Network (FCN) joint self-attention mechanism for NLOS/LOS identification. The method is validated using public datasets (European Horizon 2020 Programme project eWINE), and the proposed method is found to have the highest accuracy of about 88.24% compared to CNN, LSTM, CNN-LSTM, FCN, and LSTM-FCN [17]. However, this method is affected by the quality and quantity of the training datasets. Refs. [18,19] proposed a method that converts one-dimensional CIR data into two-dimensional images and uses deep learning networks for identification. Its accuracy is affected by image size and inefficient operation.
To address the problems in the above methods, such as manual feature extraction leads to incomplete database content of candidate classification features, difficulty in selecting appropriate thresholds in multiple scenarios, and low recognition rates of other neural network methods. The NLOS/LOS identification method based on multilayer CNN combined with CAM is proposed. In other terms, embedding CAM in the CNN module reduces the redundant information generated in feature self-extraction and improves the characterization ability of the CNN. The input layer reduces the computational complexity by replacing the two-dimensional feature map with a one-dimensional feature map. The traditional CNN is improved by adding a batch normalization (BN) layer and a Rectified Linear Unit (ReLU) between the convolutional and pooling layers to speed up the convergence. Moreover, a Global Average Pooling (GAP) layer is chosen instead of a fully connected layer to reduce the training parameters and improve the model’s generalization ability. It obtains better identification results using CNN-CAM compared with other ways. The specific details of the proposed method are described in Section 4.

3. Preliminaries

In this section, the variability of UWB-ranging performance in the NLOS/LOS environment is tested. In addition, the performance of CIR is analyzed using the IEEE802.15.4a standard channel model.

3.1. NLOS/LOS Problem Statement

In complex indoor environments, the obstacles between the transmitter and receiver result in signal propagation through multiple paths. Among them, the LOS path means that the signal propagates directly between the transmitter and the receiver. The NLOS path means that the signal reaches the receiver by reflection, diffraction, and scattering. The LOS and NLOS propagation schematic is shown in Figure 1.
It can be seen from Figure 1a that the direct physical link between devices is not obscured in the LOS environment. The distance between the devices can be accurately estimated using the propagation time of the UWB direct path signal. It can be seen from Figure 1b that the direct physical link is obscured, causing the direct path signal to be curtailed and not accurately received. The UWB signals are affected by reflection, refraction, and scattering from obstacles during propagation, resulting in the additional distance. In this case, there is a delay in signal arrival time, which leads to reduced ranging accuracy.
UWB-based Indoor Positioning System (IPS) uses distance information from different channels to calculate positioning results. To test the variability of UWB ranging performance in LOS and NLOS environments, ranging experiments were conducted in LOS/NLOS environments. The ranging error results for the LOS and NLOS environments are provided in Table 1 and Table 2, respectively. The bar graph of ranging error at different distances in LOS/NLOS environments is shown in Figure 2. In addition, the mean and standard deviation of the ranging errors for each reference distance in both the LOS and NLOS environments are the result of calculations using 192 data.
Table 1 and Table 2 and Figure 2 show that the range error between the anchor node and the target node in the LOS environment is not large, and its mean and standard deviation of the range error do not exceed 0.1198 m and 0.0497 m, respectively. However, the mean value of the range error in the NLOS environment is not less than 0.3476 m, and the standard deviation of the range error is not less than 0.0318 m. Therefore, it is necessary to perform NLOS identification before positioning to achieve better positioning results.

3.2. CIR Performance Analysis

The CIR is the sum of the received pulses obtained by evaluating the correlation between the cumulative incoming samples and the expected lead sequence [20].
As shown in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 (the red lines in Figure 4, Figure 5, Figure 6 and Figure 7 represent their respective mean values), the IEEE802.15.4a standard channel model is chosen to analyze the CIR performance in LOS and NLOS environments. Firstly, the indoor residential LOS environment (CM1) and the indoor residential NLOS environment (CM2) are selected. Secondly, the corresponding specific parameters are selected according to the channel environment [21]. In addition, the continuous pulse function of the channel is realized according to the specific parameters, and then the continuous pulse function is discretized. Finally, RMS delay spread, mean excess delay, number of effective paths with a peak within 10 dB, number of valid paths with energy greater than 85%, and average power are calculated.
From Figure 3, it can be seen that the signal attenuation is relatively slow in the LOS environment, and the peak value of the CIR waveform is high, which is because the signal can reach the receiver through a direct path. However, in the NLOS environment, the signal amplitude is relatively tiny and decays quickly due to the obstacle blockage. In the LOS environment, the RMS delay spread, and the mean excess delay are shorter than the average time required in the NLOS environment. In the case of the same number of channels, the average values of the number of effective paths with peaks within 10 dB and the number of effective paths with energy greater than 85% are smaller in the LOS environment than in the NLOS environment. A comparison of the average power attenuation curves shows that the NLOS environment takes longer than the LOS environment when the receiver receives a signal with the same attenuation power.
In summary, the performance of CIR in NLOS and LOS environments is significantly different, so the deep learning method can directly use CIR as the input vector for NLOS /LOS identification.

4. Method

This paper aims to build a real-time NLOS/LOS identification method with a high recognition rate, high environmental applicability, and low computational complexity. The CNN combined with CAM for NLOS/LOS identification method is proposed. In this section, the related theories of CNN and CAM parts are introduced, and the proposed CNN-CAM network architecture and identification steps are described in detail.

4.1. CNN Theory

CNN is a feed-forward neural network inspired by natural biological visual cognitive mechanisms [22], which performs multiple convolution and pooling operations on the input data using multiple filters to obtain high-level features inside the data [23]. The structure of One-dimensional CNN mainly consists of convolution layers, pooling layers, and fully connected layers.
The convolution layer performs convolutional operations by sliding convolution kernels, and the output of these kernel filters is usually fed into the activation function to extract features. The one-dimensional CNN formula is as in Equation (1):
y j l = f i = 1 N w i j l x i l 1 + b j l
where indicates the convolution calculation; N is the number of kernel in ( l 1 ) th layer; x i l 1 is the i feature mapping of the ( l 1 ) th layer; b j l is the bias of the j th convolution kernel of the l th layer; w is the weight; y j l is the feature map representing the output of the j th convolution of the l th layer; f ( · ) represents the nonlinear activation function.
The activation function can enhance the nonlinear expression ability of the model. In this paper, we use the ReLU function, which enables neurons with sparse activation. The expression is shown in Equation (2).
f R e L U ( x ) = max ( 0 , x )
To further reduce the training parameters, the pooling or subsampling operation is often required, usually using Max-pooling or average pooling to compress the data in the sliding region. In this paper, Max-pooling is used, and the process is represented by Equation (3):
p i l + 1 ( j ) = max ( j 1 ) ω + 1 t j ω { q i l ( t ) }
where q i l ( t ) is the value of the t th neuron corresponding to the i th feature; ω is the pooling layer width; p i l + 1 ( j ) is the ( l + 1 ) th layer neuron value.
After alternating convolutional and pooling layers several times, the pooling layers are flattened and connected to one or more fully connected layers to achieve classification.

4.2. Attention Mechanism

Attention mechanisms have been widely used in natural language processing, data prediction, hydroacoustic identification, image segmentation, etc. Compared to deep learning network architectures, the attention mechanism is a lightweight module that tunes the network parameters by generating and assigning weights and trains the network to focus on key information to improve accuracy [24,25]. In this paper, the CAM is added to the CNN, and its structure is shown in Figure 9. Firstly, the input feature maps are transformed Into two one-dimensional vectors by Global Max Pooling (GMP) and GAP, respectively. Then, they are passed through a shared Multilayer Perceptron (MLP). Finally, to obtain the weight values of the channels corresponding to the feature maps, the two output terms of the MLP are summed by channel and normalized using the activation function sigmoid.
The corresponding theoretical Equation is (4):
M c ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) ) = σ W 1 W 0 F avg c + W 1 W 0 F max c
where σ denotes the sigmoid function, both W 0 and W 1 are weights of MPL, and F avg c and F m a x c are the characteristics of GAP and GMP, respectively [26].

4.3. NLOS/LOS Identification Method Based on CNN-CAM

In this section, the advantages of the fusion of both CNN and CAM are explained, the CNN-CAM network structure is constructed, the parameters of each layer and the identification steps are described in detail, and the performance evaluation metrics are given.

4.3.1. CNN-CAM Network Architecture

The CIR of UWB can be considered a time series, and there is a correlation between the before and after data under LOS conditions. The data under NLOS conditions have apparent differences. Therefore, the CNN network is introduced into the paper, which has more advantages in learning the structural relationship between CIR data. However, traditional CNN takes the same way to convolve each channel of the feature map. In fact, different channels carry different importance of information, so processing each channel in the same way will degrade the accuracy of the network. CAM is a lightweight and universal module. It can not only assign different weights to input features to highlight important features and suppress useless feature responses but also integrates seamlessly with any CNN architecture for end-to-end training. Its advantages have been validated on different classification and detection datasets. Therefore, this paper embeds CAM in the multilayer CNN and builds the NLOS/LOS identification system based on CNN-CAM. The CNN-CAM network architecture is shown in Figure 10, and its specific parameters are shown in Table 3.
As can be seen from Figure 10 and Table 3, the CNN-CAM network architecture consists of three parts.
The first part contains three convolution layers, and each convolution module consists of a convolution layer, a BN layer, a ReLU function, and a Max-pooling layer. The input vector was the CIR data with the size 1016 × 1. Moreover, the first convolution layer uses the 4 × 1 convolution kernel with a number of 10 to perform the initial feature extraction operation. The second convolutional layer uses the 5 × 1 convolution kernel with a number of 20. The third convolutional layer uses the 3 × 1 convolution kernel with a number of 32 to mine deeper information from the output features of the upper layer. The stride size of all convolution layers is 2. The sizes of the convolution kernels used in this network’s first and second convolution layers were obtained from the literature [11,12]. In the third convolutional layer, we have used a 3 × 1 convolution kernel. This convolution kernel was obtained after conducting experiments on the effect of different sizes of convolutional kernels on the identification accuracy. In the experiment, 1 × 1, 2 × 1, 3 × 1, 4 × 1, and 5 × 1 are selected as the selection list of convolution kernel size. The BN can make the input samples become normally distributed with mean 0 and variance 1, thus solving the problem of slow learning speed due to the scattered distribution of sample features. Therefore, the paper adds BN layers after each convolutional layer to speed up the model training. And chooses ReLU as the activation function after BN layers. The pooling layer not only reduces the size of the parameter matrix but also allows filtering operations for additional noise introduced by the CIR signal under the influence of hardware circuits, transmission paths, NLOS receiving surfaces, and other factors. Therefore, the paper adds a 2 × 1 Max-pooling layer with a step size of 2 after each convolution layer.
The second part adds CAM on top of the above to further enhance the feature extraction capability of the model. The Max-pooling layer in the third convolution module is used as the input of CAM, and the GMP and GAP operations are performed, respectively. Subsequently, the two pooling layers sequentially perform a convolution of 1 × 1 of the number 8, the BN layer, the ReLU function, and the convolution of 1 × 1 of the number 32. Finally, the operations of superposition, sigmoid function, and multiplication are experienced.
The third part uses a convolution layer to transform the data dimensions and a GAP layer instead of a fully connected layer for feature integration. Finally, the softmax activation function accomplishes the identification of NLOS/LOS.

4.3.2. NLOS/LOS Identification Process

The NLOS/LOS identification process based on CNN-CAM is shown in Figure 11. The steps are as follows:
  • The obtained CIR data is divided into training sets, validation sets, and test sets in the ratio of 7:2:1.
  • Train the CNN-CAM model with the training sets and validate the performance of the trained model with the validation sets. Furthermore, the trained model is saved when the epoch is reached.
  • The trained model is tested with test sets to obtain the final NLOS/LOS identification result.
In this paper, the order of the training set is randomly disrupted with the aim of improving the robustness of the model. In addition, the Adam optimizer is used for each training period, the learning rate decay period is set to 10, and the learning rate decay is 0.5 times the original.
In addition, to evaluate the performance of the proposed model, we use four metrics: Accuracy, LOS recall, NLOS recall, and F1-score. As shown in Equations (5), (6), (7), and (8), respectively [8,9].
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l - L O S = T P T P + F N
R e c a l l - N L O S = T N T N + F P
F 1 - s c o r e = 2 T P 2 T P + F N + F P
where T P is the number of data correctly identified as LOS, T N is the number of data correctly identified as NLOS, F N is the misjudgment data for LOS, and F P is the misjudgment data for NLOS.

5. Results and Discussion

In this section, firstly, the dataset was briefly described and visualized. Secondly, experiments on model parameters are designed to determine the optimal values of learning rate and batch size. Finally, to verify the advancedness and effectiveness of the model, experiments with different structural models and different identification methods were conducted.

5.1. Visual Analysis of Datasets

The public dataset was used for experiments [27]. The data are measured in seven indoor scenarios, including office 1, office 2, a small apartment, a small workshop, a kitchen with a living room, a bedroom, and a boiler room. In addition, 3000 LOS and 3000 NLOS channel measurements were collected in each scenario. Figure 12 shows the CIR sampling points in the seven environments, and Figure 13 shows the numerical distribution of the signal characteristic parameters in LOS and NLOS environments.
From Figure 12, it can be seen that the CIR waveform is not clearly distinguished in each environment, reflecting the complexity of this dataset. From Figure 13, it can be seen that the maximum amplitude, rise time, noise standard deviation, received signal power, kurtosis, skewness, total energy, RMS delay spread, peak time, mean excess delay, peak-to-average ratio, the amplitude of the steep rise in energy have relatively high overlap. And it shows that the difference between LOS and NLOS datasets is not apparent.
In summary, the dataset selected in this paper has the test conditions to verify the proposed method.

5.2. Experiments and Results

In NLOS/LOS identification based on CNN-CAM, there are mainly three parts of the experiment. Firstly, we conducted experiments for different sizes of learning rates and training batches to select the appropriate parameters. In addition, we designed six models with different structures to verify the effectiveness of adding the CAM and the three-layer convolution module (Convolution + BN + ReLU + Max-pooling). Finally, to verify the model’s advancedness, we compare the proposed method with the state-of-the-art NLOS/LOS identification methods, such as the identification method of joint machine learning of channel parameters and the identification method of other deep learning.

5.2.1. Parameter Analysis

Model parameters play an essential role in the performance of the network. Suitable parameters not only improve the convergence speed in the training phase but also help to achieve better classification results. Different learning rates and batch sizes were selected to find the best value of learning rate and batch sizes for performance comparison, as shown in Figure 14a,b.
The learning rate refers to the magnitude of each update of the parameters. If the learning rate is too low, the convergence speed of the network will be slow. If the learning rate is too large, the optimized parameters will fluctuate repeatedly near the optimal value, making network convergence difficult. From Figure 14a, it can be seen that there is a relationship between the model’s accuracy and the learning rate, with a maximum value of 89.62% achieved at 0.001. However, the accuracy rate decreases when the learning rate decreases to 0.0001. Therefore, the best learning rate determined using the CNN-CAM model is 0.001.
When the training batch is too small, the difference in the samples leads to an extensive range of statistical characteristics of the batch, which makes the direction of gradient descent unstable, so the identification accuracy fluctuates wildly. When the training batch is too large, due to the limitation of the training set size, its number of gradient descents is too tiny to find the global optimal solution easily. At this time, the statistical characteristics of large-scale samples are considered, which do not accurately represent the direction of gradient update in the training set. From Figure 14b, it can be observed that the model achieves the highest accuracy when the training batch size is set to 64, suggesting that optimal results can be achieved by setting the model batch size to 64.

5.2.2. Performance Analysis

Various comparative experiments were designed to compare the proposed model’s performance: different structural models, identification methods for joint machine learning of feature parameters, and other deep learning identification methods. In addition, we use four metrics: Accuracy, Recall-LOS, Recall-NLOS, and F1-score to evaluate the model performance.
(a)
Comparative Experiments of Different Structural Models
To verify the validity of the proposed model, the paper constructed six models with different structures and analyzed their performance, as shown in Figure 15 and Figure 16 and Table 4, respectively.
As can be seen from Figure 15 and Figure 16 and Table 4, Model A did not add CAM, and its accuracy was the lowest. Model B consists of a layer of convolutional modules (Convolution + BN + ReLU + Max-pooling), CAM, and “Convolution + GAP”. The accuracy leveled off when the training model increased with the number of iterations. At this point, the accuracy of model B has increased by 1.83%, the LOS recall has increased by 4.05%, the NLOS recall has remained essentially unchanged, and the F1-score has increased by 2.08%, indicating that the addition of CAM can improve the identification accuracy. Model C adds a layer of convolution module (Convolution + BN + ReLU + Max-pooling) on the basis of Model B. Its accuracy is 88.40%, LOS recall is 91.38%, NLOS recall is 85.43%, F1-score is 88.74%, and the number of parameters is 4804. The proposed CNN-CAM model is based on model B with the addition of a two-layer convolution module (Convolution + BN + ReLU + Max-pooling), which achieves an accuracy of 90.00%, a LOS recall of 92.29%, an NLOS recall of 87.71%, an F1-score of 90.22%, and a parameter count of 8764. Model C and the proposed CNN-CAM model increase the number of parameters, but improve the accuracy by 2.00% and 3.60%, the LOS recall by 0.62% and 1.53%, the NLOS recall by 3.38% and 5.66%, and the F1-score by 1.77% and 3.25%, respectively, when compared to Model B. It indicates that adding the convolution module can improve the identification accuracy. Model D continues to add a layer of convolution module, which reduces the recognition rate, verifying the feasibility of using a three-layer convolution module. Model E replaces the GAP in the proposed model with a fully connected layer. Every neuron in the fully connected layer is linked to all the neurons in the previous layer, and the number of trainable parameters increases from 8764 to 70,204, which causes an overfitting phenomenon.
In summary, by analyzing the performance of the above models with different structures, it is found that the CNN-CAM model proposed in this paper ensures good fitting performance with fewer trainable parameters and the highest model accuracy, which verified the effectiveness of the CNN-CAM model proposed.
(b)
Comparison Experiments of Different Identification Methods
To verify the advancedness of the proposed CNN-CAM model, CNN-LSTM [15], CNN-SVM, and Random Forest (RF) models are selected for comparison with the model proposed in this paper. The CNN-SVM model is based on the feature vectors extracted from the CNN fully connected layer as input, and the kernel function of the SVM classifier is the radial basis function (RBF). The RF (single feature) approach takes a kurtosis feature as input and uses RF for identification. RF (multiple features) method is used to extract 12 feature parameters of CIR data, including kurtosis, skewness, total energy, RMS delay spread, peak time, mean excess delay, peak-to-average ratio, amplitude of the steep rise in energy, maximum amplitude, rise time, noise standard deviation, and received signal power, and then use RF to identify them. In addition, the paper sets the number of decision trees in the RF algorithm to 100, and the minimum number of samples of leaf nodes is set to 1. A comparison of the performance of different identification methods is shown in Table 5.
As shown in Table 5, the CNN-CAM model proposed in this paper achieves 90.00% accuracy, 92.29% LOS recall, 87.71% NLOS recall, and 90.22% F1-score. In terms of accuracy, compared to CNN-LSTM, CNN-SVM, RF (single feature), and RF (multiple features), the accuracy of the CNN-CAM model is improved by 5.06%, 3.88%, 35.48%, and 2.57%, respectively. In terms of LOS recall, CNN-CAM has a significant improvement effect compared to CNN-LSTM, CNN-SVM, RF (single feature), and RF (multiple features), with a minimum improvement of 6.62% and a maximum of 38.21%. In terms of NLOS recall, CNN-CAM outperforms CNN-LSTM, CNN-SVM, and RF (single feature) and is slightly lower than RF (multiple features). However, the use of multiple feature parameters for identification improves the NLOS recall, but the accuracy and stability are affected because there is some irrelevant and redundant information between the artificially extracted features. In terms of F1-score, CNN-CAM shows significant improvement compared to CNN-LSTM, CNN-SVM, RF (single feature) and RF (multiple features). Furthermore, the lowest improvement is 2.37%, and the highest is 36.12%. In summary, the CNN-CAM network model proposed in this paper extracts features with higher sensitivity and has obvious performance improvement effects compared to the neural network models and machine learning models in existing studies. Moreover, it is more suitable for NLOS/LOS identification.

6. Conclusions

In this paper, firstly, the features of CIR data in LOS and NLOS environments are analyzed in detail, and the factors affecting range are identified. Then, aiming at the problem of low accuracy and poor environmental adaptability of existing UWB CIR identification methods, the NLOS/LOS identification method based on multilayer CNN combined with CAM is proposed. The method takes the one-dimensional CIR signal as input, uses three groups of convolution modules (Convolution + BN + ReLU + Max-pooling) and CAM for feature self-extraction, and uses a GAP for feature integration to achieve NLOS/LOS identification. A Max-pooling layer is used to achieve the filtering operation for the extra noise introduced by the CIR signal. The overfitting phenomenon is avoided by replacing the fully connected layer with the GAP layer, which reduces the parameters by 87.52% compared to Model E. In addition, to compare the proposed model’s performance, a variety of comparative experiments were designed using public datasets: different structural models, identification methods for joint machine learning of feature parameters, and other deep learning identification methods. It is found that the proposed model has 90.00% accuracy, 92.29% LOS recall, 87.71% NLOS recall, and 90.22% F1-score, which achieves high identification results in NLOS/LOS identification of UWB and verifies the effectiveness and advancement of the model.
However, the inability to collect large datasets in real emergency scenarios limits the application scenarios of this network to some extent. Therefore, NLOS/LOS identification in the small sample case will be considered in future work.

Author Contributions

Conceptualization, J.Z. and L.H.; data curation, J.Z., Q.Y., Z.Y., J.C. and H.Z.; formal analysis, J.Z.; methodology, J.Z. and L.H.; software, J.Z.; writing-original draft, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “National Key Research and Development Program of China”, grant number “2021YFB3900800”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is not provided temporarily due to restrictions, such as privacy or ethics.

Acknowledgments

This work was supported in part by the National Key Research and Development Program of China (project: High Precision Positioning Navigation and Control Technology for Large Underground Space (No. 2021YFB3900800)).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, J.; Chen, R.; Pei, L.; Guinness, R.; Kuusniemi, H. A hybrid smartphone indoor positioning solution for mobile LBS. Sensors 2012, 12, 17208–17233. [Google Scholar] [CrossRef] [PubMed]
  2. Huang, L.; Yu, B.; Du, S.; Li, J.; Jia, H.; Bi, J. Multi-Level Fusion Indoor Positioning Technology Considering Credible Evaluation Analysis. Remote Sens. 2023, 15, 353. [Google Scholar] [CrossRef]
  3. Alarifi, A.; Al-Salman, A.; Alsaleh, M.; Alnafessah, A.; Al-Hadhrami, S.; Al-Ammar, M.A.; Al-Khalifa, H.S. Ultra wideband indoor positioning technologies: Analysis and recent advances. Sensors 2016, 16, 707. [Google Scholar] [CrossRef]
  4. Zhang, H.; Wang, Q.; Yan, C.; Xu, J.; Zhang, B. Research on UWB Indoor Positioning Algorithm under the Influence of Human Occlusion and Spatial NLOS. Remote Sens. 2022, 14, 6338. [Google Scholar] [CrossRef]
  5. Kong, Y.; Li, C.; Chen, Z.; Zhao, X. Recognition of Blocking Categories for UWB Positioning in Complex Indoor Environment. Sensors 2020, 20, 4178. [Google Scholar] [CrossRef]
  6. Yu, K.; Wen, K.; Li, Y.; Zhang, S.; Zhang, K. A novel NLOS mitigation algorithm for UWB localization in harsh indoor environments. IEEE Trans. Veh. Technol. 2018, 68, 686–699. [Google Scholar] [CrossRef]
  7. Tuchler, M.; Huber, A. An improved algorithm for UWB-bases positioning in a multi-path environment. In Proceedings of the 2006 International Zurich Seminar on Communications, Zurich, Switzerland, 22 February 2004–24 February 2006; pp. 206–209. [Google Scholar]
  8. Liu, Q.; Yin, Z.; Zhao, Y.; Wu, Z.; Wu, M. UWB LOS/NLOS identification in multiple indoor environments using deep learning methods. Phys. Commun. 2022, 52, 101695. [Google Scholar] [CrossRef]
  9. Yang, H.; Wang, Y.; Seow, C.K.; Sun, M.; Si, M.; Huang, L. UWB sensor-based indoor LOS/NLOS localization with support vector machine learning. IEEE Sens. J. 2023, 23, 2988–3004. [Google Scholar] [CrossRef]
  10. Si, M.; Wang, Y.; Siljak, H.; Seow, C.; Yang, H. A lightweight CIR-based CNN with MLP for NLOS/LOS identification in a UWB positioning system. IEEE Commun. Lett. 2023, 27, 1332–1336. [Google Scholar] [CrossRef]
  11. Güvenç, İ.; Chong, C.C.; Watanabe, F.; Inamura, H. NLOS identification and weighted least-squares localization for UWB systems using multipath channel statistics. EURASIP J. Adv. Signal Process. 2007, 2008, 271984. [Google Scholar] [CrossRef]
  12. Marano, S.; Gifford, W.M.; Wymeersch, H.; Win, M.Z. NLOS identification and mitigation for localization based on UWB experimental data. IEEE J. Sel. Areas Commun. 2010, 28, 1026–1035. [Google Scholar] [CrossRef]
  13. Li, H.; Xie, Y.; Liu, Y.; Ye, X.; Wei, Z. Research on NLOS identification method based on CIR feature parameters. China Test. 2021, 47, 20–25. [Google Scholar]
  14. Jiang, C.; Chen, S.; Chen, Y.; Liu, D.; Bo, Y. An UWB channel impulse response de-noising method for NLOS/LOS classification boosting. IEEE Commun. Lett. 2020, 24, 2513–2517. [Google Scholar] [CrossRef]
  15. Jiang, C.; Shen, J.; Chen, S.; Chen, Y.; Liu, D.; Bo, Y. UWB NLOS/LOS classification using deep learning method. IEEE Commun. Lett. 2020, 24, 2226–2230. [Google Scholar] [CrossRef]
  16. Li, J.; Deng, Z.; Wang, G. NLOS/LOS signal recognition model based on CNN-BiLSTM. In Proceedings of the China Satellite Navigation Conference (CSNC), Beijing, China, 26–28 April 2022; p. 7. [Google Scholar]
  17. Pei, Y.; Chen, R.; Li, D.; Xiao, X.; Zheng, X. FCN-Attention: A deep learning UWB NLOS/LOS classification algorithm using fully convolution neural network with self-attention mechanism. Geo-Spat. Inf. Sci. 2023, 1–20. [Google Scholar] [CrossRef]
  18. Cui, Z.; Gao, Y.; Hu, J.; Tian, S.; Cheng, J. LOS/NLOS identification for indoor UWB positioning based on Morlet wavelet transform and convolutional neural networks. IEEE Commun. Lett. 2020, 25, 879–882. [Google Scholar] [CrossRef]
  19. Wang, J.; Yu, K.; Bu, J.; Lin, Y.; Han, S. Multi-classification of UWB signal propagation channels based on one-dimensional wavelet packet analysis and CNN. IEEE Trans. Veh. Technol. 2022, 71, 8534–8547. [Google Scholar] [CrossRef]
  20. Yang, H.; Wang, Y.; Xu, S.; Huang, L. A new set of channel feature parameters for UWB indoor localization. J. Navig. Position. 2022, 10, 43–52. [Google Scholar]
  21. Molisch, A.F.; Cassioli, D.; Chong, C.C.; Emami, S.; Fort, A.; Kannan, B.; Kardeal, J.; Schantz, H.G.; Siwiak, K.; Win, M.Z. A comprehensive standardized model for ultrawideband propagation channels. IEEE Trans. Antennas Propag. 2006, 54, 3151–3166. [Google Scholar] [CrossRef]
  22. Huang, L.; Yu, B.; Li, H.; Zhang, H.; Li, S.; Zhu, R.; Li, Y. HPIPS: A high-precision indoor pedestrian positioning system fusing WiFi-RTT, MEMS, and map information. Sensors 2020, 20, 6795. [Google Scholar] [CrossRef]
  23. Lee, H.; Lee, Y.; Jung, S.-W.; Lee, S.; Oh, B.; Yang, S. Deep Learning-Based Evaluation of Ultrasound Images for Benign Skin Tumors. Sensors 2023, 23, 7374. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.S. SCA-CNN: Spatial and Channel-Wise Attention in Convolutional Networks for Image Captioning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5659–5667. [Google Scholar]
  25. Qu, J.; Tang, Z.; Zhang, L.; Zhang, Y.; Zhang, Z. Remote Sensing Small Object Detection Network Based on Attention Mechanism and Multi-Scale Feature Fusion. Remote Sens. 2023, 15, 2728. [Google Scholar] [CrossRef]
  26. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  27. Bregar, K.; Mohorčič, M. Improving indoor localization using convolutional neural networks on computationally restricted devices. IEEE Access 2018, 6, 17429–17441. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of LOS and NLOS propagation: (a) LOS paths; (b) NLOS paths.
Figure 1. Schematic diagram of LOS and NLOS propagation: (a) LOS paths; (b) NLOS paths.
Sensors 23 08552 g001
Figure 2. Graph of ranging error at different distances: (a) LOS environment; (b) NLOS environment.
Figure 2. Graph of ranging error at different distances: (a) LOS environment; (b) NLOS environment.
Sensors 23 08552 g002
Figure 3. CIR waveform in LOS/NLOS environment: (a) LOS CIR waveform; (b) NLOS CIR waveform.
Figure 3. CIR waveform in LOS/NLOS environment: (a) LOS CIR waveform; (b) NLOS CIR waveform.
Sensors 23 08552 g003
Figure 4. Root mean square (RMS) delay spread in LOS/NLOS environment: (a) LOS RMS delay spread; (b) NLOS RMS delay spread.
Figure 4. Root mean square (RMS) delay spread in LOS/NLOS environment: (a) LOS RMS delay spread; (b) NLOS RMS delay spread.
Sensors 23 08552 g004
Figure 5. Mean excess delay in LOS/NLOS environment: (a) LOS mean excess delay; (b) NLOS mean excess delay.
Figure 5. Mean excess delay in LOS/NLOS environment: (a) LOS mean excess delay; (b) NLOS mean excess delay.
Sensors 23 08552 g005
Figure 6. Number of effective paths with a peak within 10 dB in LOS/NLOS environment: (a) LOS number of effective paths with a peak within 10 dB; (b) NLOS number of effective paths with a peak within 10 dB.
Figure 6. Number of effective paths with a peak within 10 dB in LOS/NLOS environment: (a) LOS number of effective paths with a peak within 10 dB; (b) NLOS number of effective paths with a peak within 10 dB.
Sensors 23 08552 g006
Figure 7. The number of valid paths with energy greater than 85% in the LOS/NLOS environment: (a) LOS number of valid paths with energy greater than 85%; (b) NLOS number of valid paths with energy greater than 85%.
Figure 7. The number of valid paths with energy greater than 85% in the LOS/NLOS environment: (a) LOS number of valid paths with energy greater than 85%; (b) NLOS number of valid paths with energy greater than 85%.
Sensors 23 08552 g007
Figure 8. Average power decay curve in LOS/NLOS environment: (a) LOS average power; (b) NLOS average power.
Figure 8. Average power decay curve in LOS/NLOS environment: (a) LOS average power; (b) NLOS average power.
Sensors 23 08552 g008
Figure 9. CAM structure diagram.
Figure 9. CAM structure diagram.
Sensors 23 08552 g009
Figure 10. CNN-CAM network architecture.
Figure 10. CNN-CAM network architecture.
Sensors 23 08552 g010
Figure 11. NLOS/LOS identification process of CNN-CAM.
Figure 11. NLOS/LOS identification process of CNN-CAM.
Sensors 23 08552 g011
Figure 12. Schematic diagram of CIR sampling points in the seven environments (731 ns–830 ns): (a) LOS environment; (b) NLOS environment.
Figure 12. Schematic diagram of CIR sampling points in the seven environments (731 ns–830 ns): (a) LOS environment; (b) NLOS environment.
Sensors 23 08552 g012
Figure 13. Numerical distribution of signal characteristic parameters in LOS and NLOS environments (Purple color represents the portion where the number of channels are overlapped).
Figure 13. Numerical distribution of signal characteristic parameters in LOS and NLOS environments (Purple color represents the portion where the number of channels are overlapped).
Sensors 23 08552 g013aSensors 23 08552 g013b
Figure 14. Graph of the effect of parameters on model performance: (a) Model performance with different learning rates; (b) Model performance with different batch sizes.
Figure 14. Graph of the effect of parameters on model performance: (a) Model performance with different learning rates; (b) Model performance with different batch sizes.
Sensors 23 08552 g014
Figure 15. Models of different structures.
Figure 15. Models of different structures.
Sensors 23 08552 g015
Figure 16. Model performance comparison chart for different structures.
Figure 16. Model performance comparison chart for different structures.
Sensors 23 08552 g016
Table 1. LOS environment ranging error.
Table 1. LOS environment ranging error.
Reference Distance (m)Mean of Ranging Error (m)Standard Deviation of Ranging Error (m)
1.00000.07900.0253
2.00000.09540.0232
3.00000.09590.0306
4.00000.06530.0270
5.00000.10400.0343
6.00000.06040.0305
7.00000.11980.0497
8.00000.03510.0289
9.00000.03770.0322
Table 2. NLOS environment ranging error.
Table 2. NLOS environment ranging error.
Reference Distance (m)Mean of Ranging Error (m)Standard Deviation of Ranging Error (m)
1.80230.11860.0275
2.50000.23480.0318
3.35410.57470.2456
4.2720.34760.1922
5.22020.92030.6075
6.18472.08520.2191
7.1591.97410.0469
8.13943.17190.3178
9.12412.77600.2898
Table 3. CNN-CAM specific parameters table.
Table 3. CNN-CAM specific parameters table.
Network CompositionDesignationParameter
Part 1Sequence Input1016 × 1 × 1
Convolution_1 (stride)4 × 1 × 10 (2)
BN, ReLU——
Max-pooling_1 (stride)2 × 1 (2)
Convolution_2 (stride)5 × 1 × 20 (2)
BN, ReLU——
Max-pooling_2 (stride)2 × 1 (2)
Convolution_3 (stride)3 × 1 × 32 (2)
BN, ReLU——
Max-pooling_3 (stride)2 × 1 (2)
Part 2GMP, GAP——
Convolution_4/Convolution_61 × 1 × 8
BN, ReLU/BN, ReLU——
Convolution_5/Convolution_71 × 1 × 32
Part 3Convolution1 × 1 × 128
GAP——
Dropout0.5
TrainingEpoch25
Learning rate0.001
Batch size64
Table 4. Comparison table of model performance of different structures.
Table 4. Comparison table of model performance of different structures.
Model NamesNumber of ParameterAccuracy
(%)
Recall-LOS (%)Recall-NLOS (%)F1-Score
(%)
Model_A173684.5786.7182.4284.89
Model_B212486.4090.7682.0586.97
Model_C480488.4091.3885.4388.74
Model_proposed CNN-CAM876490.0092.2987.7190.22
Model_D775689.1991.9086.4889.47
Model_E70,20489.0590.5787.5289.21
Table 5. Performance comparison of different identification methods.
Table 5. Performance comparison of different identification methods.
MethodsAccuracy (%)Recall-LOS (%)Recall-NLOS (%)F1-Score (%)
CNN-LSTM84.9484.9184.9784.93
CNN-SVM86.1285.6786.5786.06
RF (single feature)54.5254.0854.9654.10
RF (multiple features)87.4385.5289.5987.85
CNN-CAM proposed90.0092.2987.7190.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Yi, Q.; Huang, L.; Yang, Z.; Cheng, J.; Zhang, H. Research on None-Line-of-Sight/Line-of-Sight Identification Method Based on Convolutional Neural Network-Channel Attention Module. Sensors 2023, 23, 8552. https://doi.org/10.3390/s23208552

AMA Style

Zhang J, Yi Q, Huang L, Yang Z, Cheng J, Zhang H. Research on None-Line-of-Sight/Line-of-Sight Identification Method Based on Convolutional Neural Network-Channel Attention Module. Sensors. 2023; 23(20):8552. https://doi.org/10.3390/s23208552

Chicago/Turabian Style

Zhang, Jingjing, Qingwu Yi, Lu Huang, Zihan Yang, Jianqiang Cheng, and Heng Zhang. 2023. "Research on None-Line-of-Sight/Line-of-Sight Identification Method Based on Convolutional Neural Network-Channel Attention Module" Sensors 23, no. 20: 8552. https://doi.org/10.3390/s23208552

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop