Next Article in Journal
Real-Time Loosely Coupled 3DMA GNSS/Doppler Measurements Integration Using a Graph Optimization and Its Performance Assessments in Urban Canyons of New York
Next Article in Special Issue
Optimization of the Trajectory, Transmit Power, and Power Splitting Ratio for Maximizing the Available Energy of a UAV-Aided SWIPT System
Previous Article in Journal
Homogeneity Characterization of Textile-Integrated Wearable Sensors based on Impedance Spectroscopy
Previous Article in Special Issue
Handover Management for Drones in Future Mobile Networks—A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAS-SEINet: A SNR-Aware Adaptive Scalable SEI Neural Network Accelerator Using Algorithm–Hardware Co-Design for High-Accuracy and Power-Efficient UAV Surveillance †

1
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Research Center of Advanced RF Chips and Systems, Nanhu Laboratory, Jiaxing 314000, China
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in 2021 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA), Zhuhai, China, 24–26 November 2021.
Sensors 2022, 22(17), 6532; https://doi.org/10.3390/s22176532
Submission received: 15 August 2022 / Revised: 23 August 2022 / Accepted: 23 August 2022 / Published: 30 August 2022
(This article belongs to the Special Issue UAV Assisted 5G and Future Wireless Networks)

Abstract

:
As a potential air control measure, RF-based surveillance is one of the most commonly used unmanned aerial vehicles (UAV) surveillance methods that exploits specific emitter identification (SEI) technology to identify captured RF signal from ground controllers to UAVs. Recently many SEI algorithms based on deep convolution neural network (DCNN) have emerged. However, there is a lack of the implementation of specific hardware. This paper proposes a high-accuracy and power-efficient hardware accelerator using an algorithm–hardware co-design for UAV surveillance. For the algorithm, we propose a scalable SEI neural network with SNR-aware adaptive precision computation. With SNR awareness and precision reconfiguration, it can adaptively switch between DCNN and binary DCNN to cope with low SNR and high SNR tasks, respectively. In addition, a short-time Fourier transform (STFT) reusing DCNN method is proposed to pre-extract feature of UAV signal. For hardware, we designed a SNR sensing engine, denoising engine, and specialized DCNN engine with hybrid-precision convolution and memory access, aiming at SEI acceleration. Finally, we validate the effectiveness of our design on a FPGA, using a public UAV dataset. Compared with a state-of-the-art algorithm, our method can achieve the highest accuracy of 99.3% and an F1 score of 99.3%. Compared with other hardware designs, our accelerator can achieve the highest power efficiency of 40.12 Gops/W and 96.52 Gops/W with INT16 precision and binary precision.

1. Introduction

With the rapid development of 5G and beyond (e.g., 6G) and wireless communication technology and the increasing complexity of the electromagnetic environment, unmanned aerial vehicles (UAVs), also known as drones, have received increasing popularity, since they offer extraordinary ability, high mobility, and low cost in aiding and improving a wireless network. For instance, an UAV can provide flexible and stable connectivity between communication devices, establish relay links [1] and cellular networks [2], assist radio localization and navigation [3], and so on [4,5,6]. While benefiting from an UAV, the misuse and “black flight” of drones is also a concern. Some civilian drones may enter restricted areas without certification, which will seriously threaten airspace security, cybersecurity, and even public safety. Despite the government’s drone regulation efforts, such as real-name registration and electronic fences, there are still some illegal drones that violate these regulatory measures.
To address the above security issue, UAV surveillance systems have been developed to detect and identify different UAVs. As one of the surveillance methods, the RF-based specific emitter identification (SEI) technology [7,8] can identify a captured RF signal from UAVs to the controllers and distinguish different UAV individuals by extracting the electromagnetic signal characteristics, known as a RF fingerprint. The RF fingerprint [9,10] is an inherent feature caused by the imperfections of a RF circuit including I/Q imbalance, phase noise, frequency offset, etc. These imperfections have unique and inevitable characteristics, which are hard to imitate. At the same time, the RF fingerprint contains rich information about UAVs, which allows RF-based surveillance to be less-constrained in many scenarios. To obtain such a RF fingerprint, SEI based on machine learning (ML) [11,12], especially a deep convolutional neural network (DCNN) [13,14,15,16,17], can automatically extract deeper features of RF signals with high accuracy, which has attracted many researchers. However, these methods have high computational complexity, requiring a domain-specific DCNN processor for acceleration. Although some DCNN processors have been proposed, most of them are used in computer vision and other fields, which are not suitable for SEI in terms of power efficiency.
In this paper, we propose a high-accuracy and power-efficient SEI accelerator using an algorithm–hardware co-design for UAV surveillance [18]. The main contributions are described as follows:
  • On the algorithm level, scalable SEI neural network with SNR-aware adaptive precision computation is proposed to deal with the UAV identification task under different SNRs. Then, a 16-bit DCNN and a binary DCNN are used for low SNR and high SNR, respectively. Two DCNNs can be adaptively switched according to the SNR estimated by the second and fourth moments (M2M4) algorithm, which can reduce the power consumption while ensuring the accuracy.
  • On the algorithm level, a short-time Fourier transform (STFT)-based feature extraction, reusing the DCNN method, is proposed to pre-extract features of a UAV signal. It allows the reuse of the convolution operators of a DCNN and reduces hardware costs. In addition, we use normalization, quantization, and denoising preprocessing methods to improve the overall accuracy.
  • On the hardware level, a DCNN engine with hybrid-precision convolution and memory access is proposed, which speeds up the computation and reduces hardware costs. The hybrid-precision convolution can be reused by the convolution, binary convolution, and STFT convolution operation. The hybrid-precision memory access can reuse the parameter storge of a DCNN and a binary DCNN.
  • On the hardware level, the specialized SNR sensing engine and denoising engine are designed for SEI. Denoising engine is responsible for denoising the RF data to reduce signal redundancy. A SNR sensing engine is responsible for estimating the SNR of the emitter signal, which determines whether we use a DCNN or abinary DCNN.
  • The rest of this paper is organized as follows: Section 2 reviews the related work and techniques; Section 3 focuses on the proposed algorithm–hardware co-design for a SEI accelerator; Section 4 describes the dataset, neural network architecture, evaluation method, test setup, and experimental results; Section 5 compares our method with other algorithms and hardware designs; and Section 6 presents the conclusions.

2. Related Works

Recently, many ML-based SEI methods have been proposed for UAV surveillance. Al-Sa’d et al. [14] first built a public RF-based UAV dataset, including the RF data of different UAVs in different flight modes, such as: off mode, on and connected mode, hovering mode, and flight and video recording mode. To confirm the feasibility of the above RF-based UAV dataset, they designed three four-layer deep neural networks (DNN), with the same architecture for detecting the presence of UAVs, the presence of UAVs and their types, and, finally, the presence of UAVs, their types, and their flight modes. The classification categories for the three tasks are 2, 4, and 10 categories, respectively. As the difficulty of the classification task gradually increases, the accuracy of the first DNN (class 2) decreases from 99.7% to 84.5% in the next DNN (class 4) and, finally, to 46.8% in the third DNN (class 10).
Al-Emadi et al. [15] used a six-layer DCNN for UAV detection and an eight-layer DCNN for UAV-type classification and flight-pattern classification. A one-dimensional convolution layer, one-dimensional pooling layer, and dense layer form their DCNN. Compared with a DNN, a DCNN can achieve better accuracy of 99.8%, 85.8%, and 59.2%, respectively.
Allahham et al. [16] proposed a data channelization preprocessing method and multi-channel one-dimensional DCNN architecture. For preprocessing, they divided the full Wi-Fi frequency spectrum (80 MHz for the 13 overlapping channels) into 8 equal-bandwidth channels via data channelization technology. For multi-channel one-dimensional DCNN, the network is a five-layer structure, with multichannel inputs that correspond to each separated spectrum channel. After the training process, their method can learn and analyze the feature of the different frequency band RF data, which improved the accuracy of three classification task by 0.2%, 8.8%, and 28.2%, respectively.
In [17], a multi-channel deep neural network with a joint feature engineering generator method is proposed by Yang et al. Unlike the multi-channel DCNN in the previous paper, their two-channel neural network extracts the features of the high-frequency and low-frequency components separately in the shallow layer and fuses them before the final, fully connected layer. For the feature-engineering generator, data truncation and the moving average filter are utilized to remove the noise effects. Separated normalization is utilized to train the neural network more easily rather than normalizing together, which prevents the smaller-valued high-frequency component from being dominated by the low-frequency component. Experiments shows the effectiveness of their method, which improves the accuracy of 10 categories to 98.2%.
Nemer [12] et al. propose a hierarchical learning approach for UAV identification and detection. Specifically, three tasks use three levels of ensemble learning classifiers in a cascaded form, which include a classifier for detecting UAVs in the first level, a classifier for detecting UAV types in the second level, and the last two classifiers for detecting Bebop and AR UAV flight patterns in the third level. The KNN and XGBoost classifiers form an ensemble classifier, with a final output that is based on the voting of the outputs from these two classifiers. The results show that their method can detect the presence of an UAV and identify the type of an UAV and the corresponding flight pattern with an average accuracy of about 99.2%.
Along with the extensive research to improve the accuracy of UAV identification algorithms, hardware implementation is also a key part of SEI-based UAV surveillance deployment. Similar to our work, Soltani et al. [19,20] designed an embedded implementation of a deep-learning-based classifier named DeepRadio for the modulation classification of RF signals. Unlike SEI, this classifier classifies the received signals into different modulation types. In their experiments, the DeepRadio successfully identifies the different modulation types of one USRP N210 with high accuracy and low power consumption.
In general, the existing algorithms focus only on the improvement of accuracy and lack algorithm–hardware co-design, which is not hardware-friendly. If such algorithm is directly applied to the hardware, it may bring high power consumption. Although some general-purpose processors (e.g., CPUs and GPUs) and ML-based processors are available, they are not specifically optimized for SEI and, therefore, cannot meet the real-time or power-efficiency requirements of state-of-the-art algorithms.

3. Proposed Algorithm–Hardware Co-Design for SEI Accelerator

The SEI-based UAV surveillance platform is shown in Figure 1. It consists of an UAV, a remote controller, a RF system, a repository, and a SEI system. RF systems such as universal software radio peripherals (USRP) can collect unique RF signals from different types of UAVs with different flight modes by passively and continuously listening to the communication between the UAV and the remote controller, which includes control command signals (controller to UAV), telemetry signals, and video signals (UAV to controller). The captured RF signals are then stored in a local database repository, and the stored data can be analyzed by a SEI system to detect the presence of the drone, the type of drone, and the flight pattern of the drone.

3.1. SNR-Aware Adaptive Scalable SEI Neural Network

For the SEI algorithm, the most conventional method based on deep learning is to use a uniform high-precision type of neural network for the classification, such as INT16, FLOAT32, and even FLOAT64 [21]. If this kind of SEI algorithm is directly applied to hardware, it will bring more power consumption in simple occasions with low precision requirements. In other deep-learning-application fields, binary DCNN has shown great advantages in low precision, which can reduce the overall hardware overhead. Here, aiming at the application of SEI, we take the advantages of binary DCNN and propose our SNR-Aware adaptive Scalable SEI neural Network (SAS-SEINet), as shown in Figure 2.
The SAS-SEINet algorithm includes the signal preprocessing, SNR-aware precision reconfiguration, and scalable SEI neural network. Signal preprocessing is responsible for normalization, quantization, denoising, and STFT. SNR-aware precision reconfiguration is performed to estimate the SNR based on the second and fourth moments (M2M4) algorithm and adjust the neural network precision according to the threshold judgment. The scalable SEI neural network, also denoted as SEI-DCNN, identifies the different emitter signals ranging with different SNRs. For low SNR, the SEI neural network is the conventional DCNN with INT16 precision, which can maintain the accuracy. For high SNR, the binary DCNN is applied to reduce power consumption, while the accuracy does not drop too much.

3.1.1. Scalable SEI-DCNN with SNR-Aware Adaptive Precision Computation

Conventional DCNN with fixed high-precision parameters brings high memory usage and power consumption. Those “simple” samples can be classified well with lower precision parameters. Especially for RF signals, signals with high SNR are easier to recognize than those with low SNR. Therefore, the precision with adaptive reconfiguration is more suitable for processing signals with different SNRs. Based on the above, we propose a scalable SEI-DCNN with SNR-aware adaptive precision computation, as shown in Figure 3. Under the control of SNR-aware precision reconfiguration, the precision of SEI neural networks can be reconfigured toward different SNR. To maintain the accuracy at low SNR, the backbone neural network is a conventional DCNN with 16-bit activation and 16-bit weight. To reduce the algorithm complexity and power consumption at high SNR, the backbone neural network is a binary DCNN with 16-bit activation and 1-bit weight.
SNR-aware precision reconfiguration consists of SNR estimation based on M2M4 and adaptive precision reconfiguration. The M2M4 method proposed in [22] successfully estimates the carrier strength and noise strength in a complex AWGN channel. Since the M2M4 does not need carrier recovery and has a wide range of SNR estimation, it is more suitable for practical applications. After the estimation of M2M4, the estimated SNR is sent to the adaptive precision reconfiguration, based on a simple judgment mechanism. If the estimated SNR is greater than a threshold, the binary DCNN will be used to process such signals with high SNR. Otherwise, the DCNN will be used to process signal with low SNR. The threshold in judgment mechanism is obtained from the experiment, which can be referred to in Section 4.5.
Specifically, the expression of SNR estimation based on M2M4 is as follows:
S N R = E s N = 2 M 2 2 M 4 M 2 2 M 2 2 M 4 = ( 1 2 ( M 4 M 2 2 ) 1 ) 1
where M2 and M4 represent the second and the fourth moments of the received signal y n . Due to the fact that the statistical average of received signals cannot be obtained directly, one general method is to approximate the statistical mean by time average, as follows:
{ M 2 1 N n = 0 N 1 | y n | 2 M 4 1 N n = 0 N 1 | y n | 4
With the increase in the number of observation data (denoted as N), the SNR value estimated by M2M4 is closer to the real value. In addition, it is found that the standard deviation of SNR estimation is less than 0.2 dB when the N is more than 2000.
For scalable SEI-DCNN, its backbone neural network is a four-layer convolutional neural network, including a STFT convolution layer, two convolutional layers, and a fully connected (FC) layer. Each of convolutional layers consists of convolution, activation function, and average pooling.
  • Convolution layer
In a conventional DCNN [23], the common operation of a convolution layer can be expressed as follows:
A n + 1 = p ( f ( A n     W n ) )
where A n + 1 denotes the (n + 1)th layer output tensor generated by the previous layer tensor A n and corresponding weight tensor W n . f denotes the activation function (e.g., sigmoid and ReLu), which introduces the non-linearity to the model. p denotes the pooling function, which compresses the activation values and removes the redundant information. Here, we use the ReLu activation and average pooling function. Standard convolution operation includes multiplication and addition operations, which occupies the majority computation of the DCNN.
Unlike DCNN, the standard convolution is replaced by binary convolution in a binary DCNN [24], as shown in Figure 4. Binary convolution uses binary weights for the convolution, which is implemented by addition and subtraction operations instead of multiplication operations. Thus, a binary DCNN greatly reduces the memory usage and computation. Specifically, the convolution operation in a binary DCNN can be transformed as follows:
A n     W n ( A n B W n ) α
where represents binary convolution without any multiplication. The binarized weight B W n is derived from W n . The scale factor α introduces a small amount of multiplication, but it will improve the overall classification accuracy. For the convolution layer, different scale factors are used to multiply the convolution results of each output channel. For the fully connected layer, the scale factor of the output neuron itself is used to multiply the result of matrix multiplication.
The optimal value of the binary weight B W n used in the binary neural network can be obtained by taking the sign of the original weight W n .
B W n = s i g n ( W n )
The optimal value of the scale factor α is obtained by averaging the sum of the absolute values of the elements w n i in the original weight tensor W n is expressed as:
α =   | w n i | n = 1 n w n l 1
  • FC layer
The FC layer [23] acts as the classifier and is usually located in the last layers of the DCNN. Unlike the previous layer that maps the initial input to the hidden space to extract features, the FC layer maps the learned hidden features to the label space. Specifically, the operation of FC layer can be expressed as:
A n + 1 = f ( A n × W n )
where × denotes the matrix multiplication operation between previous layer tensor A n and corresponding weight tensor W n . f denotes the activation function (e.g., sigmoid and softmax), which produces the score or probability of each category. Here, the activation function is available for DCNN training. During inference phase, the matrix multiplication can be replaced by a full convolution operation, and the activation function can be optionally skipped, which does not affect the final classification result.
  • STFT convolution layer
As one of the widely used preprocessing method, STFT can map a one-dimensional time-domain signal into a joint distribution of time and frequency, preserving both the time-domain and frequency-domain features of the signal. In this paper, we merged the STFT preprocessing into the DCNN as a convolution layer. The details of this method will be introduced in the following section.

3.1.2. STFT-Based Feature Extraction Reusing DCNN

Although a DCNN has powerful automatic feature-extraction capabilities, directly feeding raw data without any processing into neural network may make training difficult to converge and result in poor performance. Therefore, researchers often perform appropriate preprocessing on RF signals to improve the overall accuracy of the algorithm. In this paper, we propose a STFT reusing a DCNN method to extract the feature. STFT is implemented as a STFT convolution layer, which allows the reuse of the DCNN and reduces hardware costs.
Given a window function ω with length N and stride s, the standard STFT [25] amplitude spectrum | X s t f t ( t , f ) | of original signal x can be expressed as:
| X s t f t ( t , f ) | = | n = x [ n ] ω [ n s t ] e i 2 π n N f |
To make the computation of STFT more convenient, we derive Equation (8) as:
| X s t f t ( t , f ) | = | ( x [ t ] ω [ t ] e i 2 π t N f ) S T F T   k e r n e l w i t h   s t r i d e   s |  
where ⊗ denotes the convolution operation between x and the STFT kernel. With the Euler formula, the complex STFT kernel can be split into real and imaginary parts:
K r e a l ( t , f ) = ω [ t ] cos ( 2 π t f / N ) K i m a g ( t , f ) = ω [ t ] sin ( 2 π t f / N )
By substituting Equation (10) into Equation (9), we can obtain:
| X s t f t ( t , f ) | = | x [ t ] K r e a l ( t , f ) + i × x [ t ] K i m a g ( t , f ) |
where denotes the convolution operation. It can be seen that the formulation of the STFT amplitude spectrum can be expressed as the combination of two one-dimensional convolutions. In other words, we can use two one-dimensional convolutions with the K r e a l and K i m a g kernels to compute the STFT amplitude spectrum instead, which reuses our convolution operator of DCNN and is easier to implement.
In addition to STFT, normalization, quantization, and denoising preprocessing methods are also used to process raw data. These methods are described in detail as follows:
  • Normalization and quantization
Normalization [26] is the process of scaling individual samples to have a unit norm. It makes all samples have the same range and facilitates convergence of training. In this paper, we use min–max normalization to scale the RF samples to the range of [−1, 1]. Specifically, our normalization is formulated as follows:
X m i n m a x = 2 ( x x m i n ) x m a x x m i n 1
where x and X m i n m a x are the input RF data and the output quantized data, respectively. The minimum and maximum values of x are denoted by x m i n and x m a x .
For quantization, we utilized the INT16 quantization based on Kullback–Leibler divergence (KLD) to quantize the input data. Such a method [27] attempts to approximate the original numerical distribution of FP32 with that of INT16, which ensures the accuracy of the network after quantization and facilitates deployment of FPGA. Specifically, the quantization is formulated as follows:
Z q u a n t i z a t i o n = ( 2 15 1 ) × c l i p ( z , | T | , | T | ) | T | .  
where z and Z q u a n t i z a t i o n denote the original FP32 data and the quantized INT16 data, respectively. |T| is the saturation threshold of quantization, and c l i p ( z , | T | , | T | ) is the function that truncates the original data z to the range of [− | T | , | T | ]. Generally, the threshold |T| is less than the maximum of | z m i n | and | z max | .
Instead of directly mapping the range of [ z m i n , z m a x ] to [− ( 2 15 1 ) , ( 2 15 1 ) ], KLD-based quantization truncates the values outside the ± | T | and maps the range of [− | T | , | T | ] to [− ( 2 15 1 ) , ( 2 15 1 ) ], which prevents the accuracy from being affected by the abnormal z m a x and z m i n . Additionally, the quantization tries to adjust the threshold |T| to approximate the distribution of the INT16 data to the original distribution of the FP32 data. The distribution similarity can be measured by the KLD. The smaller the KLD value, the more similar the two distributions are, and the best threshold |T| can be obtained when the KLD value is minimal.
  • Denoising
Raw RF data often include long segments of noise, as shown in Figure 5, and such segments may dominate the entire signal, bringing confusion to the training of model. To reduce the effect of background noise, denoising is an effective means to separate the UAV signal from the noise. Since the amplitude of noise is much smaller than that of the signal, we utilize a short-time energy detection to extract the signal [28]. Specifically, the short-time energy E m can be calculated as follows:
E m = n = ( x [ n ] ω [ m n ] ) 2
where x and ω are the input data and window function, respectively. When the short-time energy E m in the window is higher than a certain threshold, we can judge that there is a valid signal in the window, otherwise there is noise.

3.2. Reconfigurable Hybrid-Precision SEI Hardware Accelerator

The architecture of the proposed SEI hardware accelerator is shown in Figure 6. The program control unit (PCU) reads the user instructions (e.g., SNR threshold, DCNN structure) from the instruction buffer and controls the SEI acceleration to achieve programmability. The data mover and DDR controller are responsible for the data transmission between on-chip and off-chip. Once the storage capacity of the on-chip buffer exceeds, the off-chip DDR can be used for storage. The score comparator obtains the results of the last layer of output neurons and produces the final classification result.
Dedicated to the SEI algorithm, we design three core computation modules including a denoising engine, SNR sensing engine, and DCNN engine. The denoising engine interacts with the window buffers that store the window function data and is responsible for denoising the RF data to reduce signal redundancy. The SNR sensing engine is responsible for estimating the SNR of the emitter signal, which determines whether we use a DCNN or a binary DCNN. For DCNN processing, the double feature buffers, hybrid-precision weight buffers, bias buffers, and scaling factor buffers are used to store the input–output feature map data and trained parameters of DCNN. Under the condition of sufficient storage capacity, these buffers can store a limited-size DCNN fully on-chip without having to go off-chip, which reduces the latency and power consumption. In particular, the hybrid-precision weight buffers can store both 16-bit weight or binary weight in a compact-storage strategy, and, thus, a larger neural network can be stored in binary weight mode. Besides, the DCNN engine includes multiple processing units (PU). Each PU is composed of a hybrid-precision CONV, ReLu, and pooling module. Instead of designing two computation units separately, the hybrid-precision CONV can compute a 16-bit convolution or binary convolution in a multiplexed manner. In addition, the DCNN engine can process multiple output feature maps or multiple input feature maps with two parallel modes, respectively, according to the different characteristics of convolution and full connection (FC) computing.

3.2.1. DCNN Engine with Hybrid-Precision Convolution and Memory Access

The core processing module of DCNN engine is shown in Figure 7, including multiple PUs with a hybrid-precision CONV, ReLu, and pooling modules. For a hybrid-precision CONV module, both standard convolutional computation and binary convolutional computation can be supported. Taking 3 × 3 convolution as an example, in the standard convolution mode, nine groups of multipliers will be selected by MUX2 to calculate nine times of multiplication with the feature maps and weights in the convolution. In the binary convolution mode, nine MUX1 will be selected instead of the multiplication operation. The positive and negative of the binary weight will determine whether the feature maps are reversed. After MUX2 and before DEMUX1, both modes will share the same computation units. The adder tree adds nine groups of 16-bit data each time and produces a 16-bit result for the convolution accumulation unit (CAU). The convolution results of each feature map channel will be accumulated on the CAU, and the final convolution result of the output feature map channel will be obtained after summing with bias in standard convolution mode or scaling with the scaling factor in the binary convolution mode. For ReLu module, it receives the result after the hybrid-precision CONV module and outputs the non-negative value after MUX4, which is judged by a comparator. For the pooling module, our DCNN engine currently only support average pooling. It caches and accumulates the results after the ReLu module with the buffer. When the arbiter judges that the configured pooling length is reached, the accumulated result goes through the shifter for a shifted division operation.
In addition to a hybrid-precision convolution, we also implement a hybrid-precision memory access strategy. As shown in Figure 8, it can be seen how weights are stored in the hybrid-precision weight buffer. Taking weights of 5 × 5 × 2 × 2 size as an example, each small square represents each weight, and four large squares form a collection of all the weight data. Ri, Cj, CIk, and COt represent the weight data of row i, column j, input channel k, and output channel t, respectively. (Ri, Cj, CIk, COt) represents the weight data of row i, column j, input channel k, and output channel t.
For demonstration, all the weight data can be divided into different color blocks, according to every 16 steps. As can be seen from the table of Figure 8, the area of each kind of color needs at most 16 addresses for weight data storage and 99 addresses for all the weight data storage while storing 16-bit weights. However, only one address is required for the area of each kind of color, and six addresses are required in total while storing binary weights. The advantage of this compact storage method is that larger neural networks and more weights can be stored in binary weight mode, compared with 16-bit weight mode. In addition, multiple binary weights can be read out in parallel, which reduces read time.

3.2.2. Denoising Engine and SNR Sensing Engine

In the process of inference, the normalization and quantization are generally done off-chip. Here, we only discuss the implementation of denoising engine and signal sensing engine.
The structure of the denoising engine is shown in Figure 9. It consists of a multiplier, a squarer, an accumulator, a comparator, a multiplexer, and a buffer. The adder, buffer, and arbiter together form an accumulator. According to Equation (14) in the denoising algorithm, the input data are first multiplied with the window function in the multiplier. The square of its multiplication result is then accumulated in the accumulator. When the number of multiplication results is up to the window length, the short-time energy is obtained by the accumulator. In the comparator, the obtained short-time energy will be compared with a certain threshold. If the energy is above that certain threshold, the input data are considered as a valid signal, and the input data cached in the buffer is output via the multiplexer. Otherwise, the input data are considered as noise and the multiplexer selects zero data for output.
The structure of the signal sensing engine is shown in Figure 10. It consists of a divider, a lookup table, a comparator, two accumulators, and three squarers. According to Equation (1) in the SNR-aware precision reconfiguration algorithm, the SNR value will first be calculated by M2M4 estimation. It can be observed that the estimated SNR can be expressed as a function of an independent variable of ( M 4 M 2 2 ) . This function can be implemented by a lookup table, which saves computational costs. Therefore, we only need to calculate M 2 2 and M 4 to obtain the final estimated SNR value. For the calculation of M 2 2 , it is obtained by a squarer, an accumulator, and another squarer. For the calculation of M 4 , it is obtained by one shared squarer, another squarer, and an accumulator. After computing M 2 2 and M 4 , the estimated SNR can be inferred from the divider and the lookup table. In the comparator, the estimated SNR is compared with a certain threshold value, and a decision signal is generated. If the estimated SNR is above a certain threshold, the input data are considered as a high SNR signal, and the decision signal is pulled up. Otherwise, the input data are considered as a low SNR signal, and the decision signal is pulled down.

4. Experiments and Results

To validate the proposed DCNN processor, we have implemented it using a Zynq-7045 FPGA board.

4.1. Dataset

During our experiments, the publicly available UAV dataset [14] was chosen to validate our algorithm–hardware codesign performance, which is also convenient for comparison with existing work. Table 1 shows the composition of this dataset. In this dataset, a total of 227 segments of time-domain RF data were recorded, which can be classified into 10 types. One type is 10.25 s of background noise, and the other nine types are 5.25 s of RF data from three UAVs (AR, Bebop, and Phantom) in different flight modes, which include on and connected mode; hovering mode; flying mode; and flying with video recording mode. More details can be found in the article [29]. During training, we adopted the K-fold cross-validation method to randomly divide the dataset into 10 non-overlapping folds, of which 9 folds are used for training, and the remaining fold is used for testing. This process will be repeated 10 times by us, to evaluate the entire dataset.

4.2. SEI-DCNN Network Architecture

We use a 4-layer network architecture named SEI-DCNN for UAV identification. The detailed network structure of SEI-DCNN is shown in Table 2. Please note, we implemented the SEI-DCNN on our accelerator for demonstration, but different DCNNs can be implemented on our proposed DCNN engine by changing its user instructions.

4.3. Evaluation Method

We evaluate our proposed method from the three aspects, including accuracy, F1 score, and power efficiency.
  • Accuracy and F1 score
In the classification task, there are generally four classification cases including true positive (TP), false positive (FP), true negative (TN), and false negative (FN) [30]. TP indicates the number of positive samples accurately predicted as positive. FP indicates the number of negative samples incorrectly predicted as positive. TN indicates the number of negative samples accurately predicted as negative. FN indicates the number of positive samples incorrectly predicted as negative. Based on the above, we use accuracy and F1 score to evaluate the performance of our SEI algorithm as follow:
A c c u r a c y = T P + T N T P + T N + F P + F N
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l                                 { P r e c i s i o n = T P T P + F P               R e c a l l = T P T P + F N
The accuracy focuses on the proportion of correct samples to all test samples, while the F1 score focuses on precision and recall.
  • Power Efficiency
As one of the important and effective metric, power efficiency [31] is used to evaluate the performance of hardware, which considers not only speed but also power consumption. Specifically, GOPs/w is used to quantify power efficiency as follow:
G O P s / w = G O P s P w
where P w indicates the power consumption that can be measured in watts by the power measurement tool of the software. GOPs can be used as a measure of speed, indicating the number of giga operations per second when the algorithm is running with specific hardware.

4.4. Test Setup

Figure 11 shows a test setup before the experiment. The bitstream of the hardware design is downloaded to the FPGA in advance via the serial port. The PC is responsible for configuring the SEI-DCNN on FPGA and sending the local offline RF samples to the FPGA. After the neural network completes the identification, the returned results are printed on the display screen of PC.

4.5. Experimental Results

By counting the prediction results of DCNN for each category, we can obtain the confusion matrix of SEI-DCNN with FLOAT32 precision, as shown in Figure 12. The green cells and red cells represent the correctly and incorrectly classified samples, respectively. The yellow cells on the left and top side represent the F1 scores for each class. The gray cells on the right and bottom side represent the recall and precision value, respectively. According to the yellow blocks and gray blocks, we can obtain the average F1 scores in the orange cell and overall accuracy in the blue cell. Finally, both the classification accuracy and F1 score of our method are 99.3%.
To evaluate the performance of our algorithm under different SNRs, we added noise to the original dataset to simulate the real situation. Figure 13 shows the classification accuracy under different SNR. We compare the accuracy of DCNN with INT16 precision and binary DCNN in the SNR range of [−5 dB, 30 dB]. The step size of the SNR interval is 5. It can be seen from the figure that the accuracy of DCNN is more than 5% higher than that of binary DCNN in the SNR range of [−5 dB, 20 dB). With the increase in SNR, the accuracy of binary DCNN in the SNR range of [20 dB, 30 dB] gradually approaches the accuracy of DCNN with INT16 precision. Especially, the accuracy of both neural networks can reach more than 97% at 30 dB, but binary DCNN has greater advantages in computing and storage. Based on the above analysis, we set 20 dB as our threshold in our SNR sensing engine to determine which DCNN is used.
Table 3 shows the detailed power consumption obtained from Vivado software, including static power and dynamic power. In the low SNR condition of [−5 dB, 15 dB], the power consumption of the FPGA is about 1280 mW with 16-bit weight mode. In the high SNR condition of [20 dB, 30 dB], the power consumption of the FPGA is about 610 mW with binary weight mode.
By using 8 equally divided test samples under the low SNR range of [−5 dB, 15 dB] and high SNR range of [20 dB,30 dB], we can obtain the average accuracy and power efficiency of three methods, as shown in Figure 14. In the low SNR condition, the average accuracy of our method is 87%, which is 16% higher than that of binary DCNN. In the high SNR condition, our method has a higher power efficiency of 96.52 Gops/W compared with DCNN.

5. Comparison with Other Methods

For the algorithm, we compared our algorithm with four papers that use the same dataset in their implementations. As shown in Table 4, our algorithm can achieve an accuracy of 99.3% and a F1 score of 99.3% with FLOAT32 precision, which is better than those of other algorithms. Besides, our algorithm can achieve an accuracy of 98.5% and an accuracy of 97.5% with INT16 precision and binary precision, respectively, which is still high among these algorithms.
For hardware, we compared our accelerator with four designs including CPU, GPU, and FPGAs, as shown in Table 5. Compared with CPU- and FPGA-based designs, our design has higher computational performance. Although our computational performance is not as good as the GPU due to fewer computing resources, our design achieved the highest power efficiency of 40.12 Gops/W and 96.52 Gops/W with INT16 precision and binary precision, respectively.

6. Conclusions

In this work, we have proposed a SEI hardware accelerator with a SAS-SEINet algorithm co-designed for UAV surveillance, which has been implemented on a Zynq-7045 FPGA board. In terms of the algorithm, we propose a SAS-SEINet including signal preprocessing, a SNR-aware precision reconfiguration, and a scalable SEI neural network. In terms of hardware, a SNR sensing engine, denoising engine, and specialized DCNN engine with hybrid-precision convolution and memory access are designed for SAS-SEINet acceleration. The final results show that the accuracy of 99.3% and the F1 score of 99.3% are the best among the state-of-the-art algorithms. The power efficiency of 40.12 Gops/W and 96.52 Gops/W can be achieved with INT16 precision and binary precision, respectively, which are the highest compared with the other hardware designs.

Author Contributions

Conceptualization, J.G.; methodology, J.G., A.H. and J.Z.; investigation, Z.K., Z.Q., Z.Y., R.Y. and Y.W.; writing—original draft preparation, J.G.; writing—review and editing, J.G., H.S. and J.Z.; supervision, H.S. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was jointly supported by NSAF (No. U2030204) and National Natural Science Foundation of China (No. 62074026).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available online at https://data.mendeley.com/datasets/f4c2b4n755/1 (accessed on 1 November 2020).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
RFRadio Frequency
SEISpecific Emitter Identification
MLMachine Learning
DCNNDeep Convolution Neural Network
DNNDeep Neural Network
UHFUltra-High Frequency
USRPUniversal Software Radio Peripheral
STFTShort-time Fourier Transform
SNRSignal-to-Noise Ratio
KLDKullback-Leibler Divergence
M2M4Second and Fourth Moments
FPGAField Programmable Gate Array
CPUCentral Processing Unit
GPUGraphic Processing Unit

References

  1. Zeng, Y.; Zhang, R.; Lim, T.J. Wireless Communications with Unmanned Aerial Vehicles: Opportunities and Challenges. IEEE Commun. Mag. 2016, 54, 36–42. [Google Scholar] [CrossRef]
  2. Sekander, S.; Tabassum, H.; Hossain, E. Multi-Tier Drone Architecture for 5G/B5G Cellular Networks: Challenges, Trends, and Prospects. IEEE Commun. Mag. 2018, 56, 96–103. [Google Scholar] [CrossRef]
  3. Horapong, K.; Chandrucka, D.; Montree, N.; Buaon, P. Design and Use of “Drone” to Support the Radio Navigation Aids Flight Inspection. In Proceedings of the 2017 IEEE/AIAA 36th Digital Avionics Systems Conference (DASC), St. Petersburg, FL, USA, 17–21 September 2017; pp. 1–6. [Google Scholar]
  4. Bernardini, A.; Mangiatordi, F.; Pallotti, E.; Capodiferro, L. Drone Detection by Acoustic Signature Identification. Electron. Imaging 2017, 2017, 60–64. [Google Scholar] [CrossRef]
  5. Hua, X.; Ono, Y.; Peng, L.; Cheng, Y.; Wang, H. Target Detection Within Nonhomogeneous Clutter Via Total Bregman Divergence-Based Matrix Information Geometry Detectors. IEEE Trans. Signal Process. 2021, 69, 4326–4340. [Google Scholar] [CrossRef]
  6. Unlu, E.; Zenou, E.; Riviere, N.; Dupouy, P.-E. Deep Learning-Based Strategies for the Detection and Tracking of Drones Using Several Cameras. IPSJ Trans. Comput. Vis. Appl. 2019, 11, 7. [Google Scholar] [CrossRef]
  7. Taha, B.; Shoufan, A. Machine Learning-Based Drone Detection and Classification: State-of-the-Art in Research. IEEE Access 2019, 7, 138669–138682. [Google Scholar] [CrossRef]
  8. Nguyen, P.; Truong, H.; Ravindranathan, M.; Nguyen, A.; Han, R.; Vu, T. Cost-Effective and Passive RF-Based Drone Presence Detection and Characterization. GetMobile Mob. Comput. Commun. 2018, 21, 30–34. [Google Scholar] [CrossRef]
  9. Sankhe, K.; Belgiovine, M.; Zhou, F.; Angioloni, L.; Restuccia, F.; D’Oro, S.; Melodia, T.; Ioannidis, S.; Chowdhury, K. No Radio Left Behind: Radio Fingerprinting Through Deep Learning of Physical-Layer Hardware Impairments. IEEE Trans. Cogn. Commun. Netw. 2020, 6, 165–178. [Google Scholar] [CrossRef]
  10. Sankhe, K.; Belgiovine, M.; Zhou, F.; Riyaz, S.; Ioannidis, S.; Chowdhury, K. ORACLE: Optimized Radio Classification through Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Communications, INFOCOM 2019, Paris, France, 29 April–2 May 2019; pp. 370–378. [Google Scholar]
  11. Ezuma, M.; Erden, F.; Kumar Anjinappa, C.; Ozdemir, O.; Guvenc, I. Detection and Classification of UAVs Using RF Fingerprints in the Presence of Wi-Fi and Bluetooth Interference. IEEE Open J. Commun. Soc. 2020, 1, 60–76. [Google Scholar] [CrossRef]
  12. Nemer, I.; Sheltami, T.; Ahmad, I.; Yasar, A.U.-H.; Abdeen, M.A.R. RF-Based UAV Detection and Identification Using Hierarchical Learning Approach. Sensors 2021, 21, 1947. [Google Scholar] [CrossRef] [PubMed]
  13. Huynh-The, T.; Pham, Q.-V.; Nguyen, T.-V.; Costa, D.B.D.; Kim, D.-S. RF-UAVNet: High-Performance Convolutional Network for RF-Based Drone Surveillance Systems. IEEE Access 2022, 10, 49696–49707. [Google Scholar] [CrossRef]
  14. Al-Sa’d, M.F.; Al-Ali, A.; Mohamed, A.; Khattab, T.; Erbad, A. RF-Based Drone Detection and Identification Using Deep Learning Approaches: An Initiative towards a Large Open Source Drone Database. Future Gener. Comput. Syst. 2019, 100, 86–97. [Google Scholar] [CrossRef]
  15. Al-Emadi, S.; Al-Senaid, F. Drone Detection Approach Based on Radio-Frequency Using Convolutional Neural Network. In Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT), Doha, Qatar, 2–5 February 2020; pp. 29–34. [Google Scholar]
  16. Allahham, M.S.; Khattab, T.; Mohamed, A. Deep Learning for RF-Based Drone Detection and Identification: A Multi-Channel 1-D Convolutional Neural Networks Approach. In Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT), Doha, Qatar, 2–5 February 2020; pp. 112–117. [Google Scholar]
  17. Yang, S.; Luo, Y.; Miao, W.; Ge, C.; Sun, W.; Luo, C. RF Signal-Based UAV Detection and Mode Classification: A Joint Feature Engineering Generator and Multi-Channel Deep Neural Network Approach. Entropy 2021, 23, 1678. [Google Scholar] [CrossRef] [PubMed]
  18. Gan, J.; Wang, S.; Qu, Z.; Hu, A.; Yang, Z.; Yang, R.; Wang, Y.; Shao, H.; Zhou, J. A Power-Efficient Specific Emitter Identification Hardware Accelerator With SNR-Aware Adaptive Precision Reconfiguration. In Proceedings of the 2021 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA), Zhuhai, China, 24–26 November 2021; pp. 157–158. [Google Scholar]
  19. Soltani, S.; Sagduyu, Y.E.; Hasan, R.; Davaslioglu, K.; Deng, H.; Erpek, T. Real-Time and Embedded Deep Learning on FPGA for RF Signal Classification. In Proceedings of the 2019 IEEE Military Communications Conference (MILCOM), MILCOM 2019, Norfolk, VA, USA, 12–14 November 2019; pp. 1–6. [Google Scholar]
  20. Soltani, S.; Sagduyu, Y.E.; Hasan, R.; Davaslioglu, K.; Deng, H.; Erpek, T. Real-Time Experimentation of Deep Learning-Based RF Signal Classifier on FPGA. In Proceedings of the 2019 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), Newark, NJ, USA, 11–14 November 2019; pp. 1–2. [Google Scholar]
  21. Camus, V.; Mei, L.; Enz, C.; Verhelst, M. Review and Benchmarking of Precision-Scalable Multiply-Accumulate Unit Architectures for Embedded Neural-Network Processing. IEEE J. Emerg. Sel. Top. Circuits Syst. 2019, 9, 697–711. [Google Scholar] [CrossRef]
  22. Xu, H.; Wei, G.; Zhu, J. A Novel SNR Estimation Algorithm for OFDM. In Proceedings of the 2005 IEEE 61st Vehicular Technology Conference, Stockholm, Sweden, 30 May–1 June 2005; pp. 3063–3067. [Google Scholar]
  23. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional Networks and Applications in Vision. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 30 May–2 June 2010; pp. 253–256. [Google Scholar]
  24. Rastegari, M.; Ordonez, V.; Redmon, J.; Farhadi, A. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. arXiv 2016, arXiv:1603.05279. [Google Scholar]
  25. Ye, W.; Chen, H.; Li, B. Using an End-to-End Convolutional Network on Radar Signal for Human Activity Classification. IEEE Sens. J. 2019, 19, 12244–12252. [Google Scholar] [CrossRef]
  26. Patro, S.G.K.; Sahu, K.K. Normalization: A Preprocessing Stage. arXiv 2015, arXiv:1503.06462. [Google Scholar] [CrossRef]
  27. Choukroun, Y.; Kravchik, E.; Yang, F.; Kisilev, P. Low-Bit Quantization of Neural Networks for Efficient Inference. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 3009–3018. [Google Scholar]
  28. Ge, C.; Sun, W.; Luo, C.; Yang, S.; Luo, Y. For RF Signal-Based UAV States Recognition, Is Pre-Processing Still Important at the Era of Deep Learning? In Proceedings of the 2021 7th International Conference on Computer and Communications (ICCC), Chengdu, China, 10–13 December 2021. [Google Scholar]
  29. Allahham, M.S.; Al-Sa’d, M.F.; Al-Ali, A.; Mohamed, A.; Khattab, T.; Erbad, A. DroneRF Dataset: A Dataset of Drones for RF-Based Detection, Classification and Identification. Data Brief 2019, 26, 104313. [Google Scholar] [CrossRef] [PubMed]
  30. Mo, Y.; Huang, J.; Qian, G. Deep Learning Approach to UAV Detection and Classification by Using Compressively Sensed RF Signal. Sensors 2022, 22, 3072. [Google Scholar] [CrossRef] [PubMed]
  31. Tu, F.; Yin, S.; Ouyang, P.; Tang, S.; Liu, L.; Wei, S. Deep Convolutional Neural Network Architecture with Reconfigurable Computation Patterns. IEEE Trans. Very Large Scale Integr. Syst. 2017, 25, 2220–2233. [Google Scholar] [CrossRef]
  32. Emad, A.; Mohamed, H.; Farid, A.; Hassan, M.; Sayed, R.; Aboushady, H.; Mostafa, H. Deep Learning Modulation Recognition for RF Spectrum Monitoring. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems (ISCAS), Daegu, Korea, 22–28 May 2021; pp. 1–5. [Google Scholar]
Figure 1. Block diagram of UAV surveillance platform.
Figure 1. Block diagram of UAV surveillance platform.
Sensors 22 06532 g001
Figure 2. SNR-Aware adaptive Scalable SEI neural Network.
Figure 2. SNR-Aware adaptive Scalable SEI neural Network.
Sensors 22 06532 g002
Figure 3. Scalable SEI-DCNN with SNR-aware adaptive precision computation.
Figure 3. Scalable SEI-DCNN with SNR-aware adaptive precision computation.
Sensors 22 06532 g003
Figure 4. The comparison between standard convolution and binary convolution.
Figure 4. The comparison between standard convolution and binary convolution.
Sensors 22 06532 g004
Figure 5. The noise in raw RF data.
Figure 5. The noise in raw RF data.
Sensors 22 06532 g005
Figure 6. The architecture of proposed SEI accelerator.
Figure 6. The architecture of proposed SEI accelerator.
Sensors 22 06532 g006
Figure 7. The core processing module of DCNN engine.
Figure 7. The core processing module of DCNN engine.
Sensors 22 06532 g007
Figure 8. Hybrid-precision memory access.
Figure 8. Hybrid-precision memory access.
Sensors 22 06532 g008
Figure 9. Denoising engine.
Figure 9. Denoising engine.
Sensors 22 06532 g009
Figure 10. Signal sensing engine.
Figure 10. Signal sensing engine.
Sensors 22 06532 g010
Figure 11. Test setup.
Figure 11. Test setup.
Sensors 22 06532 g011
Figure 12. The confusion matrix of SEI-DCNN with FLOAT32 precision.
Figure 12. The confusion matrix of SEI-DCNN with FLOAT32 precision.
Sensors 22 06532 g012
Figure 13. Classification accuracy under different SNR.
Figure 13. Classification accuracy under different SNR.
Sensors 22 06532 g013
Figure 14. The histogram of average accuracy and power efficiency under the high SNR and low SNR.
Figure 14. The histogram of average accuracy and power efficiency under the high SNR and low SNR.
Sensors 22 06532 g014
Table 1. Composition of UAV dataset.
Table 1. Composition of UAV dataset.
UAVLabelType-10SegmentsSamplesRatio
Bebop2On and connected21420 × 1069.25%
3Hovering21420 × 1069.25%
4Flying21420 × 1069.25%
5Flying with video recording21420 × 1069.25%
AR6On and connected21420 × 1069.25%
7Hovering21420 × 1069.25%
8Flying21420 × 1069.25%
9Flying with video recording18360 × 1067.93%
Phantom10On and connected21420 × 1069.25%
No UAV1Background noise41820 × 10618.06%
Table 2. SEI-DCNN network architecture.
Table 2. SEI-DCNN network architecture.
LayerEmbedded StructureOutput ShapeParameter
0Input Layer(None, 16,384, 1)-
1STFT-Conv(None, 128, 128, 1)65,280
2Conv2D + average pooling(None, 64, 64, 32)320
3Conv2D + average pooling(None, 16, 16, 64)18,496
4Flatten + FC(None, 10)163,850
Table 3. The power consumption on FPGA.
Table 3. The power consumption on FPGA.
SNR (dB)Weight
Precision
Static PowerDynamic PowerTotal
Power
[−5, 15]INT16245 mW1035 mW1280 mW
[20, 30]Binary245 mW365 mW610 mW
Table 4. The comparison of our algorithm with existing algorithms.
Table 4. The comparison of our algorithm with existing algorithms.
MethodAccuracyF1 Score
[14]46.8%43.0%
[15]59.2%55.1%
[13]95.4%95.0%
[17]98.4%98.3%
[12]99.2%99.1%
Ours99.3% 199.3% 1
98.5% 298.4% 2
97.5% 397.3% 3
1 is obtained with FLOAT32 precision, 2 is obtained with INT16 precision, and 3 is obtained with binary precision.
Table 5. The comparison of our hardware design with other designs.
Table 5. The comparison of our hardware design with other designs.
DesignPlatformWeight
Precision
Complexity
(Mop)
Time
(µs)
Computational
Performance
(Gops)
Chip
Power
(W)
Power
Efficiency
(Gops/W)
CPUINTEL
I5-6500
FP3264.234610.9113.9330.290.46
GPUNVIDIA GTX 1660FP3264.23301.21213.2414.0215.21
MILCOM2019 [19]
(FPGA)
XCZU9EGINT160.3624.0015.181.1513.17
ISCAS2021 [32]
(FPGA)
ZCU104INT160.8926.7833.080.8538.92
Ours
(FPGA)
XC7Z045INT16
Binary
64.23
64.23
1246.12
1090.77
51.54
58.88
1.28
0.61
40.27
96.52
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gan, J.; Hu, A.; Kang, Z.; Qu, Z.; Yang, Z.; Yang, R.; Wang, Y.; Shao, H.; Zhou, J. SAS-SEINet: A SNR-Aware Adaptive Scalable SEI Neural Network Accelerator Using Algorithm–Hardware Co-Design for High-Accuracy and Power-Efficient UAV Surveillance. Sensors 2022, 22, 6532. https://doi.org/10.3390/s22176532

AMA Style

Gan J, Hu A, Kang Z, Qu Z, Yang Z, Yang R, Wang Y, Shao H, Zhou J. SAS-SEINet: A SNR-Aware Adaptive Scalable SEI Neural Network Accelerator Using Algorithm–Hardware Co-Design for High-Accuracy and Power-Efficient UAV Surveillance. Sensors. 2022; 22(17):6532. https://doi.org/10.3390/s22176532

Chicago/Turabian Style

Gan, Jiayan, Ang Hu, Ziyi Kang, Zhipeng Qu, Zhanxiang Yang, Rui Yang, Yibing Wang, Huaizong Shao, and Jun Zhou. 2022. "SAS-SEINet: A SNR-Aware Adaptive Scalable SEI Neural Network Accelerator Using Algorithm–Hardware Co-Design for High-Accuracy and Power-Efficient UAV Surveillance" Sensors 22, no. 17: 6532. https://doi.org/10.3390/s22176532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop