Next Article in Journal
Enhancing Security and Trust in Internet of Things through Meshtastic Protocol Utilising Low-Range Technology
Next Article in Special Issue
Interference Situational Aware Beam Pointing Optimization for Dense LEO Satellite Communication System
Previous Article in Journal
IoT-Based Intrusion Detection System Using New Hybrid Deep Learning Algorithm
Previous Article in Special Issue
Trajectory and Phase Shift Optimization for RIS-Equipped UAV in FSO Communications with Atmospheric and Pointing Error Loss
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UAV Swarm Centroid Tracking for Edge Computing Applications Using GRU-Assisted Multi-Model Filtering

1
Department of Electronic and Optical Engineering, Space Engineering University, Beijing 101416, China
2
Key Laboratory of Intelligent Space TT&C and Operation, Ministry of Education, Beijing 101416, China
3
School of Space Information, Space Engineering University, Beijing 101416, China
4
China Unicom Research Institute, Beijing 100176, China
5
714 Research Institute of China State Shipbuilding Corporation Limited, Beijing 100176, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(6), 1054; https://doi.org/10.3390/electronics13061054
Submission received: 22 February 2024 / Revised: 8 March 2024 / Accepted: 8 March 2024 / Published: 12 March 2024
(This article belongs to the Special Issue Mobile Networking: Latest Advances and Prospects)

Abstract

:
When an unmanned aerial vehicles (UAV) swarm is used for edge computing, and high-speed data transmission is required, accurate tracking of the UAV swarm’s centroid is of great significance for the acquisition and synchronization of signal demodulation. Accurate centroid tracking can also be applied to accurate communication beamforming and angle tracking, bringing about a reception gain. Group target tracking (GTT) offers a suitable framework for tracking the centroids of UAV swarms. GTT typically involves accurate modeling of target maneuvering behavior and effective state filtering. However, conventional coordinate-uncoupled maneuver models and multi-model filtering methods encounter difficulties in accurately tracking highly maneuverable UAVs. To address this, an innovative approach known as 3DCDM-based GRU-MM is introduced for tracking the maneuvering centroid of a UAV swarm. This method employs a multi-model filtering technique assisted by a gated recurrent unit (GRU) network based on a suitable 3D coordinate-coupled dynamic model. The proposed dynamic model represents the centroid’s tangential load, normal load, and roll angle as random processes, from which a nine-dimensional unscented Kalman filter is derived. A GRU is utilized to update the model weights of the multi-model filtering. Additionally, a smoothing-differencing module is presented to extract the maneuvering features from position observations affected by measurement noise. The resulting GRU-MM method achieved a classification accuracy of 99.73%, surpassing that of the traditional IMM algorithm based on the same model. Furthermore, our proposed 3DCDM-based GRU-MM method outperformed the Singer-KF and 3DCDM-based IMM-EKF in terms of the RMSE for position estimation, which provides a basis for further edge computing.

1. Introduction

UAV swarms have emerged as a promising edge computing platform capable of supporting high-speed data transmission by leveraging nodes with strong communication capabilities [1,2,3]. In contrast to conventional satellite systems, which depend on supplementary data like predicted ephemeris information to support phase-locked loop (PLL) operations, highly dynamic UAV swarms face scarcity of easily accessible auxiliary information [4]. This absence of information poses challenges in achieving signal alignment, compensating for Doppler frequency offsets, and other critical operations. To address these challenges, this paper introduces a novel scenario wherein a coordinated UAV swarm is employed for edge computing applications. The proposed approach focuses on ensuring that a high-throughput UAV responsible for ground transmission remains positioned at the centroid of the swarm. Trajectory tracking of the UAV swarm’s centroid using ground-based radar observations is employed to aid with rapid acquisition and robust tracking of the PLL in ground signal demodulation. Moreover, accurate tracking of the UAV swarm’s centroid enables precise beamforming and directionality, resulting in enhanced receiver gain and improved signal reception rates. By emphasizing the tracking of the UAV swarm’s centroid, this research aims to provide solutions specifically tailored to this aspect, contributing to advancements in tracking methodologies within UAV swarm technology.
Some research has been conducted on UAV tracking based on UAV attitude, anti-drone technologies, and remote sensing imagery [5,6,7]. The theory of group target tracking (GTT) provides a flexible framework for effectively monitoring UAV swarms and enabling efficient calculation [8]. By dividing received measurements into groups based on specific criteria, GTT allows for estimating and predicting swarm trajectories and states from a group perspective [9]. The group’s state encompasses both the motion state of the group centroid and the external shape of the group distribution, making accurate tracking of the motion state crucial for the effectiveness of tracking and interception systems. As a result, tracking of maneuvering UAV swarms to achieve efficient edge computing capabilities has gained prominence.
Previous research on tracking maneuvering centroids has primarily focused on motion modeling and filtering algorithms. However, the success of filtering algorithms in achieving the desired tracking accuracy depends on the modeling of the target’s motion, for which prior information is often insufficient [10]. Many models that try to capture the target’s motion, such as the classic Singer [11], current statistical [12], and Jerk models [13], all assume that the target’s acceleration or rate of acceleration change is independent in a decoupled coordinate system. However, most target maneuvers exhibit coupling across different coordinates, particularly from the observer’s coordinate system [10]. To overcome this issue, some studies have adopted dynamic models of target characteristics to represent the control quantity generated by target maneuvers as a random process, such as the two-dimensional model [14], three-dimensional turning model [15,16,17], and flight dynamics model [18,19]. In [19], the authors considered the influence of aircraft attitude angles on target acceleration and utilized a 16-dimensional state extended Kalman filter, significantly improving the prediction accuracy of the target’s state. However, the required attitude angles were unobservable with only position measurements, and the set parameters were limited to specific scenarios, thereby restricting the practical application of this model in tracking aircraft targets.
The various models mentioned above, along with models with different parameters, correspond to distinct maneuvering modes of the target. To address the issue of motion pattern uncertainty in target tracking, a multi-model (MM) algorithm was proposed in [20]. The MM algorithms are classified into three generations: autonomous [21], cooperative [22,23], and variable structure MM methods [24,25,26,27,28,29]. Two challenges arise when employing methods that incorporate multiple filtered states. The first challenge involves using an incomplete set of models to estimate the state, while the second pertains to tracking delays during model transitions [30]. In the IMM algorithm, the transition probability matrix contains information from past models over a period of time. As a result, the IMM algorithm exhibits significant delays in response to model transitions, leading to substantial tracking errors during the initial phase of such transitions [30,31,32,33]. To mitigate model estimation delays, several approaches have been proposed. Among them, the fuzzy-tuned IMM algorithm adjusts the transition probabilities between models [34], while an adaptive neural fuzzy inference system (ANFIS) is used for mobile robot localization [35]. Deep learning techniques have been introduced into the Bayesian filtering framework due to their capabilities of nonlinear mapping and feature extraction using data-driven approaches. Recurrent neural networks (RNNs) excel at extracting features from sequential data, making them widely employed for learning motion models [36,37], noise parameters [38], and Kalman gains [39] in state estimation tasks. When dealing with switching systems, RNNs are combined with IMM methods to learn mode probabilities, which facilitate trajectory estimation, prediction, and human posture estimation. For switching systems, RNNs are combined with IMM methods to learn mode probabilities, which are used to estimate and predict trajectories and estimate human posture [40]. In combination with multi-model tracking, the long short-term memory (LSTM) network is used to replace the transition probability matrix to realize the function of model judgment [41]. Currently, such methods have not been applied to tracking a UAV swarm’s maneuvering centroid.
In summary, tracking a UAV swarm’s maneuvering centroid presents two main challenges in the process of edge computing. One is to achieve a more realistic dynamic modeling of the UAV swarm’s centroid when only position observations are available. Another challenge is accurate centroid tracking in the presence of various maneuverable scenarios and potential model transitions at any time. Thus, we propose multi-model filtering assisted by a GRU network based on a novel 3D coordinate-coupled dynamic model. The key contributions of this paper can be summarized as follows:
  • Firstly, we propose an approach for tracking the centroid of a UAV swarm by utilizing a 3D coordinate-coupled dynamic model. In this model, the control quantity is considered a random process, and an unscented Kalman filter is employed.
  • Secondly, we introduce a novel GRU-based network that extracts features from position observations. This enables quick and accurate detection of the maneuvering mode. With the detected features, a multi-model filtering method is applied to track the UAV swarm’s centroid.
  • Finally, we conduct simulations and experiments to validate the feasibility and superior performance of the proposed algorithm compared with traditional approaches.
The following sections provide further details. Section 2 discusses centroid tracking in GTT and presents the overall framework of our algorithm. Section 3 focuses on the state filtering for the UAV swarm’s centroid based on the dynamic modeling of UAV swarms. Section 4 describes how the GRU network assists with filtering the UAV swarm’s centroid under different maneuvering modes. Section 5 and Section 6 demonstrate the effectiveness of our tracking method through simulated UAV swarm trajectories and flight experiments. Finally, we conclude our work in Section 7.

2. Background Information

The proposed method in this paper aims to improve the accuracy of UAV swarm centroid tracking by utilizing a three-dimensional coordinate-coupled dynamic model and multi-model filtering assisted by a GRU network. The improved tracking results can be used for ground communication by airborne nodes in UAV swarm-based edge computing. Therefore, this section provides the key background information to familiarize readers with the working model of the system and the main principles involved.
To begin with, we introduce the scenario of UAV swarm-based edge computing and centroid tracking in it. Moreover, the working principle of centroid tracking in GTT is introduced. Subsequently, we outline the fundamental principle of modeling unknown control quantities as random processes in target tracking. Finally, we present a comprehensive framework that outlines the proposed tracking method.

2.1. Centroid Tracking in UAV Swarm-Based Edge Computing

As illustrated in Figure 1, the edge computing architecture based on UAV swarms aims to provide edge computing services for various types of users by integrating a communication node (called a centroid UAV) and multiple UAVs responsible for edge computing. The centroid UAV plays a crucial role in the overall communication and coordination of the management cluster, with the primary responsibility of efficiently distributing and relaying data between edge computing nodes, ensuring high-speed transmission and seamless connectivity of different tasks. At the same time, the centroid UAV is also responsible for high-speed data transmission to the communication gateway. The relative positions between the UAV swarm colonies are pre-programmed so that the main communication node is located at the center of mass of the UAV swarm, which serves as the dynamic anchor point of the whole UAV swarm. Different from satellite communication using ephemeris prediction to assist with the function of PLL operaitons, the trajectory information of a UAV swarm’s centroid is used to assist the acquisition, synchronization, and Doppler frequency offset of ground station signal demodulation. Since the centroid UAV is placed in the centroid position of the UAV swarm in advance, the tracking result of the centroid of the UAV swarm is the track of the centroid UAV. It is not necessary to recalculate the phase center when the trajectory information is used to demodulate the signal. The communication gateway performs two main steps: tracking the centroid of the dUAV swarm by using radar observations and the GTT algorithm. The results of centroid tracking are used to assist with the acquisition, synchronization, beamforming, and angle tracking of the receiving antenna in digital signal demodulation.
A group target, as defined in [42], comprises multiple targets that meet specific criteria. The characteristics of a group encompass its label, centroid motion state, and overall shape, as outlined in [9]. GTT involves five key steps to effectively track UAV swarms: measurement partitioning, group data association, centroid state estimation, group shape estimation, and the detection of group spawning and combination:
1. Measurement partitioning: This step involves dividing the received measurements into groups based on specific criteria. The criteria can be the proximity, similarity in motion characteristics, or any other relevant factors that define a group.
2. Group data association: In this step, the measurements are associated with their corresponding groups. This can be accomplished using various techniques such as nearest neighbor association or probabilistic data association.
3. Centroid state estimation: Once the measurements are associated with their groups, the motion state of the group centroid is estimated. This includes estimating the position, velocity, and other relevant parameters of the centroid.
4. Group shape estimation: In addition to tracking the centroid state, GTT also aims to estimate the external shape of the group distribution. This helps with understanding the spatial arrangement and dynamics of the swarm.
5. Detection of group spawning and combination: UAV swarms are dynamic systems where individual UAVs can join or leave the swarm, and multiple smaller swarms can combine to form a larger swarm. Detecting such events is important for accurate tracking and prediction of swarm behavior.
By following these steps, GTT provides a comprehensive framework for tracking maneuvering UAV swarms and enabling efficient edge computing capabilities. Under predefined criteria, measurement partitioning divides the incoming measurement data into several sets, each representing a distinct group of mobile edge computing (MEC). Group data association links an unlabeled set to the corresponding group track. Shape estimation is utilized to estimate the group’s shape and attitude parameters. The spawning and combination of groups are recorded based on the movement state and shape overlap of the groups, influencing the outcomes (creation or loss of groups) in data association. Group motion state estimation determines the motion state of the group’s centroid using measurements from the same group. The initial step in this process involves obtaining centroid observations from multiple measurements by utilizing established schemes not discussed in this paper. Subsequently, the motion state of the group is estimated through dynamic modeling and nonlinear filtering of the group’s centroid. Centroid tracking, in a strict sense, encompasses both centroid state estimation and group data association. However, this paper solely focuses on the centroid tracking of a single group. The predicted state generated by centroid tracking significantly impacts the accuracy of group association and the spawning and combination of groups, serving as a critical factor in the overall functionality of the tracking system’s edge computing.

2.2. Stochastic Process Modeling of the Control Volume

The unexpected maneuvers of the object are a result of unknown time-varying state controls for the observer. When tracking maneuvering targets, Singer initially suggested modeling the acceleration of maneuvering targets as a zero-mean first-order stationary Markov process, where the acceleration value at a given time depends on the value at neighboring times. By considering the symmetry and attenuation of the correlation function of a stationary random process, the time correlation function R a τ with the offset integral τ for the maneuvering acceleration a t in Singer’s model can be expressed as shown in Equation (1):
R a τ = E a t a t + τ = σ a 2 e α τ
where σ a 2 denotes the variance of acceleration and α is the maneuvering frequency. In practical tracking applications, because different values of α adapt to different maneuvering scenarios, it is necessary to preset the value according to prior information before filtering. The continuous-time equation of a t can be expressed as shown in Equation (2):
a ˙ ( t ) = α · a t + w t
where w ( t ) represents Gaussian white noise with a variance of 2 α · σ a 2 , and the detailed derivation can be found in [11]. The state space for the continuous-time model of one-dimensional target maneuvering is formulated as shown in Equations (3) and (4):
x ˙ ( t ) x ¨ ( t ) x ( t ) = 0 1 0 0 0 1 0 0 α x ( t ) x ˙ ( t ) x ¨ ( t ) + 0 0 1 w ( t )
x ˙ ( t ) x ¨ ( t ) x ( t ) = F x ( t ) x ˙ ( t ) x ¨ ( t ) + G w ( t )
where x ( t ) , x ˙ ( t ) , x ¨ ( t ) , and x ( t ) denote the position, velocity, acceleration, and first derivative of acceleration, respectively. Upon discretization with a sampling period of T, the state transition equation for a one-dimensional maneuver is expressed as shown in Equation (5):
X k + 1 = F k X k + W k = 1 T ( α T 1 + e α T ) / α 2 0 1 ( 1 e α T ) / α 0 0 e α T X k + W k
where F k , W k , and X k represent the state transition matrix, state noise vector, and state vector at time k and X k + 1 represents the state vector at time k + 1 . The state noise covariance matrix Q k in the Kalman filter can be expressed as shown in Equation (6):
Q k = E W k W k T
where W k can be expressed as shown in Equation (7):
W k = k T ( k + 1 ) T e F [ ( k + 1 ) T τ ] G w ( τ ) d τ
The covariance matrix for state noise in a more intricate model is fundamentally derived in accordance with Equation (6). Furthermore, the acceleration of maneuvering, which represents unknown controls, can be modeled as a time-dependent second-order model and a “current” statistical model. When dealing with a target that lacks prior information, it is necessary to model the acceleration of each dimension of the target as independent random processes within the coordinate system where the observer’s point of origin is located. This modeling is both essential and feasible in such scenarios. However, in practical scenarios, each control quantity of the UAV swarm’s centroid, generated by maneuvers such as thrust acceleration in the direction of velocity extension, pitch angle, roll angle, and deflection angle control, are independent of one another in the body coordinate system. These controls are better suited to be modeled as independent random processes within the body coordinate system. As a result, this paper derives a new state model and filtering process for tracking the UAV swarm’s centroid based on the dynamic characteristics of the UAV swarm’s centroid.

2.3. Framework of the Proposed Tracking Method

Drawing inspiration from the concepts of the MM algorithm and the utilization of deep learning as demonstrated in [18,41], this paper proposes the utilization of a maneuvering model estimation network to enhance the multi-model centroid tracking of UAV swarms. A noteworthy distinction from [18] lies in the derivation of maneuvering filtering, which exhibits significant differences based on the dynamics model of UAVs. Additionally, due to variations in maneuver types and the essential features that necessitate extraction, certain segments of the network architecture have been redesigned. The tracking framework is divided into four sequential steps:
(Step 1) Initialize filters: The given fused state estimation x ^ k 1 k 1 and the error covariance matrix P k 1 k 1 at time k 1 are updated to ensure their dimensions are appropriate for each filter. Then, the modified fusion state, error covariance matrix, and centroid observations z k at time k are used as inputs for each filter.
(Step 2) State filtering based on various maneuvering models: The filters employing N maneuvering models produce N state estimations x ^ k k 1 , x ^ k k 2 , , x ^ k k N and an error covariance matrix P k k 1 , P k k 2 , , P k k N at time k.
(Step 3) Model estimation using deep learning: Based on observations of different time epochs z k , z k 1 , , z k m , the maneuvering model estimation network determines the current maneuvering mode of the UAV swarm’s centroid. It outputs weights μ 1 , μ 2 , , μ N corresponding to different maneuver modes.
(Step 4) Fusing the related states and covariance of multiple models: In (Step 2), the outputs of multiple maneuvering filters at time k are combined by a weighted summation with the weights obtained from (Step 3). The fused states x ^ k k and the error covariance matrix P k k are also used as inputs at time k + 1 .

3. The Centroid’s UKF with the Proposed 3D Coordinate-Coupled Dynamic Model

At first, drawing inspiration from [43], we represent the tangential load, normal load, and roll angle as factors that affect the three-dimensional position, velocity, pitch angle, and yaw angle of the centroid within a zero-mean first-order stationary Markov decision process. Following this, we design a nine-dimensional unscented Kalman filter, taking into account the aforementioned constraints to effectively track the centroid of the UAV swarm using only position observations, thereby enabling efficient edge computing capabilities.

3.1. Modeling Description of the UAV Swarm’s Centroid

As depicted in Figure 2, the reference origin in the Cartesian system is set as the observer, and the position of the UAV swarm can be expressed as ( x c , y c , z c ) . The velocity of the center point relative to the origin of the coordinate system can be expressed as a scalar v c . The pitch angle θ is defined as the angle between the velocity direction and the XOY plane, while the yaw angle ψ is the angle between the projection of the velocity direction onto the XOY plane and the X axis. It should be noted that “pitch” and “yaw” in this paper describe the trajectory characteristics and should not be confused with the vehicle attitude angles described in [19]:
x ˙ c = v c cos θ cos ψ y ˙ c = v c cos θ sin ψ z ˙ c = v c sin θ v ˙ c = g ( n x sin θ ) θ ˙ = g v ( n z cos ϕ cos θ ) ψ ˙ = g n z   sin ϕ v   cos θ
The state of the centroid is influenced by three control factors: the tangential load n x , normal load n z , and roll angle ϕ . The velocity v c is impacted by the decomposition of thrust into the velocity extension component and the gravitational component. The tangential acceleration of the centroid resulting from the thrust components in the direction of extended motion is denoted as n x g , where g represents the gravitational acceleration constant and n x is the dimensionless tangential load. The pitch and yaw angle indicate the track’s direction and curvature. The control inputs affecting θ and ψ include the thrust component perpendicular to the velocity direction and the roll angle determining this component’s direction. The normal acceleration of the centroid due to the thrust component perpendicular to the motion direction is defined as n z g , with n z representing the dimensionless normal load. The angle between the normal acceleration and the vertical plane of velocity corresponds to the roll angle ϕ . Equation (8) illustrates the relationships governing the states and control variables.
Until it is accurately modeled, all information about the control quantities is unknown to the observer. In addition, changes in the three control quantities are largely independent and uncoupled from each other when maneuvering a swarm of UAVs. Thus, it is appropriate to treat n x , n z , and ϕ as distinct random processes, as outlined previously. In this study, we adopt Singer’s model assumption, positing that n x , n z , and ϕ follow zero-mean first-order stationary Markov decision processes. The continuous-time equations describing these processes can be formulated as shown in Equation (9):
n ˙ z t = α n z n z t + w n z ( t ) n ˙ x t = α n x n x t + w n x ( t ) ϕ ˙ t = α ϕ ϕ t + w ϕ ( t )
where w n z ( t ) , w n x ( t ) , and w ϕ ( t ) follow the additive white Gaussian noise with variance σ n z 2 , σ n x 2 , and σ ϕ 2 , respectively, while α n z , α n x and α ϕ are the maneuvering frequencies of n x , n z , and ϕ , respectively. In addition, w n z ( t ) , w n x ( t ) , and w ϕ ( t ) are zero-mean Gaussian white noise, and their variances are defined as 2 α n z σ n z 2 , 2 α n x σ n z 2 , and 2 α n z σ ϕ 2 , respectively. Also, by assuming independence among these control parameters, we can effectively model the dynamics of the UAV swarm during maneuvers.

3.2. Nine-Dimensional UKF of the Centroid State

Integrating the dynamical equations governing the centroid of the swarm with the differential equations concerning the control volume results in a nine-dimensional state vector x , as represented in Equation (10):
x = x c , y c , z c , v c , θ , ψ , n z , n x , ϕ T .
This comprehensive state vector encapsulates the essential variables necessary for modeling the UAV swarm’s behavior and interactions during flight maneuvers. By combining these fundamental dynamics, we aim to enhance our understanding of the system’s complex behaviors and optimize the control strategies for improved performance and efficiency.
The continuous-time state space of x can be expressed as shown in Equation (11):
x ˙ ( t ) = f ( x ( t ) ) + Gw ( t ) = v c ( t ) cos θ ( t ) cos ψ ( t ) v c ( t ) cos θ ( t ) sin ψ ( t ) v c ( t ) sin θ ( t ) g ( n x ( t ) sin θ ( t ) ) ( g n z ( t ) cos ϕ ( t ) g cos θ ( t ) ) / v c ( t ) g n z ( t ) sin ϕ ( t ) / v c ( t ) cos θ ( t ) α n z n z ( t ) α n x n x ( t ) α ϕ ϕ ( t ) + 0 0 0 0 0 0 w n z ( t ) w n x ( t ) w ϕ ( t )
In the realm of tracking nonlinear state equations, common approaches involve the utilization of extended Kalman filters, unscented Kalman filters, or particle filters. In our study, we opted for the unscented Kalman Ffilter (UKF) due to its robustness in handling nonlinear state estimation tasks compared with the extended Kalman filter (EKF). The UKF offers a reliable solution while maintaining a relatively straightforward implementation, making it a preferred choice for our analysis. Additionally, the computational burden associated with the UKF’s iterative process is notably lighter than that of the particle filter, enhancing computational efficiency.
When considering a sampling interval of T S and accounting for state noise, the discrete-time state equation can be effectively computed as described in Equation (12):
x k = x k 1 + x ˙ T S = x c , k 1 + v c , k 1 T S cos θ k 1 cos ψ k 1 y c , k 1 + v c , k 1 T S cos θ k 1 sin ψ k 1 z c , k 1 + v c , k 1 T S sin θ k 1 v c , k 1 + g T S ( n x , k 1 sin θ k 1 ) θ k 1 + g T S ( n z , k 1 cos ϕ k 1 cos θ k 1 ) / v c , k 1 ψ k 1 + g T S n z , k 1 sin ϕ k 1 / v c , k 1 cos θ k 1 n z , k 1 α n z n z , k 1 T S n x , k 1 α n x n x , k 1 T S ϕ k 1 α ϕ ϕ k 1 T S + w k 1
The observation z k , composed of radar measurements (range R, azimuth A, and elevation E), can be expressed as shown in Equation (13), considering measurement errors:
z k = R k A k E k = h ( x k ) + r k = x c , k 2 + y c , k 2 + z c , k 2 arctan ( y c , k / x c , k ) arctan z c , k / x c , k 2 + y c , k 2 + r R , k r A , k r E , k
where the measurement errors vector r k is additive white noise and its error covariance matrix R k can be expressed as shown in Equation (14):
R k = diag ( σ R 2 , σ A 2 , σ E 2 )
where σ R 2 , σ A 2 , and σ E 2 are the variances of the observation errors of the range R, azimuth A, and elevation E, respectively. The computation process of the unscented Kalman filter based on the above dynamic model is as follows:
  • Initialization
    • Initialize the state estimate x ^ 1 | 1 and the error covariance matrix P 1 | 1 . The initialization of x ^ 1 | 1 can be expressed as shown in Equation (15):
      x ^ c , 0 | 0 = R 0 cos A 0 cos E 0 y ^ c , 0 | 0 = R 0 sin A 0 cos E 0 z ^ c , 0 | 0 = R 0 sin E 0 x ^ c , 1 | 1 = R 1 cos A 1 cos E 1 y ^ c , 1 | 1 = R 1 sin A 1 cos E 1 z ^ c , 1 | 1 = R 1 sin E 1 v ^ c , 1 | 1 = ( x ^ c , 1 | 1 , y ^ c , 1 | 1 , z ^ c , 1 | 1 ) ( x ^ c , 0 | 0 , y ^ c , 0 | 0 , z ^ c , 0 | 0 ) 2 / T S θ ^ c , 1 | 1 = arctan z ^ c , 1 | 1 z ^ c , 0 | 0 / ( x ^ c , 1 | 1 , y ^ c , 1 | 1 ) ( x ^ c , 0 | 0 , y ^ c , 0 | 0 ) 2 ψ ^ c , 1 | 1 = arctan y ^ c , 1 | 1 y ^ c , 0 | 0 / x ^ c , 1 | 1 x ^ c , 0 | 0 n ^ z , 1 | 1 = 0 n ^ x , 1 | 1 = 0 ϕ ^ 1 | 1 = 0
      The initialization of the matrix P 1 | 1 follows as shown in Equation (16):
      P 1 | 1 = E [ ε 1 | 1 ε 1 | 1 T ] = E [ ( x 1 x ^ 1 | 1 ) ( x 1 x ^ 1 | 1 ) T ]
      where ε 1 | 1 represents the estimated error of x ^ 1 | 1 .
    • Set the initial values for the process noise covariance matrix Q as shown in Equation (17) and the measurement noise covariance matrix R :
      Q = diag ( σ n z 2 , σ n x 2 , σ ϕ 2 )
2.
Augmentation
  • Augment the state vector x k with the process noise vector w n z , k , w n x , k , w ϕ , k T to form an augmented state vector x a as shown in Equation (18):
    x a = x c , k , y c , k , z c , k , v c , k , θ k , ψ k , n z , k , n x , k , ϕ k , w n z , k , w n x , k , w ϕ , k T
  • Augment the state covariance matrix P k | k with the process noise covariance matrix Q to form an augmented covariance matrix P a as shown in Equation (19):
    P a = P k | k 0 0 Q
3.
Sigma Point Generation
  • Compute the weight coefficients of the sigma points. The mean weight coefficients follow as shown in Equations (20) and (21):
    w m , 0 = λ λ + n a
    w m , i = w c , i = 1 2 ( λ + n a ) , i = 1 , 2 , , 2 n a
    where n a represents the dimension of the augmented state vector and λ is a tunable parameter typically set to 3 n a . The covariance weight coefficients can be expressed as shown in Equations (22) and (23):
    w c , 0 = λ λ + n a + ( 1 α 2 + β )
    w c , i = 1 2 ( λ + n a ) , i = 1 , 2 , , 2 n a
    where α and β are tunable parameters usually set to 0.5 and 2, respectively.
  • Generate sigma points from the augmented state vector x a and the corresponding covariance matrix P a using the unscented transformation as expressed in Equation (24):
    X k | k ( i ) = x a i = 0 x a + [ ( n a + λ ) P a ] i i = 1 , 2 , , n a x a [ ( n a + λ ) P a ] i n a i = n a + 1 , , 2 n a
    where [ · ] i denotes the ith column of a matrix and · represents the square root of a matrix. By linearly transforming the augmented state vector x a and combining it with the weight coefficients, we obtain a set of sigma points that represent the state distribution.
4.
State Prediction
  • Propagate the sigma points through the nonlinear state transition function f ( x k ) to obtain predicted sigma points X k + 1 | k ( i ) as expressed in Equation (25):
    X k + 1 | k ( i ) = f ( X k | k ( i ) )
  • Estimate the predicted state x ^ k + 1 | k and the predicted covariance matrix P k + 1 | k based on the predicted sigma points as expressed in Equations (26) and (27):
    x ^ k + 1 | k = i = 0 2 n a w m , i X k + 1 | k ( i )
    P k + 1 | k = i = 0 2 n a w c , i x ^ k + 1 | k X k + 1 | k ( i ) x ^ k + 1 | k X k + 1 | k ( i ) T
5.
Measurement Prediction
  • Map the predicted sigma points X k + 1 | k ( i ) through the measurement function h ( x k ) to obtain the predicted measurement sigma points Z k + 1 | k ( i ) as expressed in Equation (28):
    Z k + 1 | k ( i ) = h ( X k + 1 | k ( i ) )
  • Estimate the predicted measurement z ^ k + 1 | k and the innovation covariance matrix S k based on the predicted measurement sigma points as expressed in Equations (29) and (30):
    z ^ k + 1 | k = i = 0 2 n a w m , i Z k + 1 | k ( i )
    S k + 1 = R k + 1 + i = 0 2 n a w c , i z ^ k + 1 | k Z k + 1 | k ( i ) z ^ k + 1 | k Z k + 1 | k ( i ) T
6.
Update
  • Compute the cross-covariance matrix P xz , k + 1 between the state and measurement and the Kalman gain K k + 1 as expressed in Equations (31) and (32):
    P xz , k + 1 = i = 0 2 n a w c , i ( X k + 1 | k ( i ) x ^ k + 1 | k ) ( Z k + 1 | k ( i ) z ^ k + 1 | k ) T
    K k + 1 = P xz , k + 1 S k + 1 1
  • Update the state estimate x ^ k + 1 and the state covariance matrix P k + 1 using the actual measurement z k as expressed in Equations (33) and (34):
    x ^ k + 1 = x ^ k + 1 | k + K k + 1 ( z k + 1 z ^ k + 1 | k )
    P k + 1 = P k + 1 | k K k + 1 S k + 1 K k + 1 T
7.
Iterate
  • Repeat steps 3–6 for each new measurement.

4. GRU-Assisted Multi-Model Filtering

In this section, we will focus on the development of a GRU-based network that allows for the extraction of features from position observations. This network plays a crucial role in enabling swift and precise detection of the maneuvering mode. Once the features have been detected, we employ a multi-model filtering approach to effectively track the centroid of the UAV swarm and achieve edge computing.

4.1. Prior Knowledge

The aircraft’s motion is composed of seven basic motion modes, as described in [43]: maintaining the original state of flight, maximum acceleration straight flight, maximum overload left turn, maximum overload right turn, maximum overload climb, maximum overload dive, and maximum deceleration flight. These maneuvers exhibit significant differences in normal load, tangential load, and roll angle. Based on these characteristics, the seven basic maneuvering modes are further divided into four categories.
Mode 1 corresponds to maintaining the original state of flight, where the normal load n z , tangential load n x , and roll angle ϕ remain relatively constant over a period of time.
Mode 2 corresponds to maximum acceleration straight flight and maximum deceleration flight, where n x changes significantly while n z and ϕ change insignificantly.
Mode 3 corresponds to maximum overload climb and maximum overload dive, where n z and n x change significantly while ϕ changes insignificantly.
Mode 4 corresponds to maximum overload left turn and maximum overload right turn, where n z and ϕ change significantly while n x changes insignificantly.
To track the motion of the centroid as it switches between these modes, we employ multi-model filtering. The model transitions are determined based on historical observations of the centroid. The focus of these maneuvering frequency combinations is to differentiate between different types of maneuvers. The optimal values for the maneuvering frequencies may vary from those obtained empirically in this study. Further analysis of the track characteristics of UAV swarms can lead to optimization of the maneuvering frequency values. Although this aspect is not discussed in this work, the methodology of analysis can be found in [11].

4.2. GRU-Based Maneuver Estimation Network

The gated Rrecurrent unit (GRU)-based approach is another variant of recurrent neural networks (RNNs) that is widely used for learning features from sequence data [44]. In this paper, the GRU network is employed to estimate the model probability and replace the calculation of the model transition probability in multi-model filtering. Similar to LSTM, a GRU also utilizes gates to adaptively control the flow of information.
At each time step k, the GRU network maintains a hidden state h s k and a cell state c s k . The cell state consists of two gates: the reset gate r k and the update gate z u k . These gates, defined by point-wise sigmoid functions σ g ( · ) , determine how much previous information should be forgotten and how much new information should be kept, while input k represents the input at time step k. The calculation of the hidden state at time k can be summarized by Equation (35):
z u k = σ g ( W z · h k 1 , input k + b z ) r k = σ g ( W r · h k 1 , input k + b r ) h ˜ k = tanh ( W h · [ r k h s k 1 , x k ] + b h ) h s k = ( 1 z u k ) h s k 1 + z u k h ˜ s k
In the above equations, W z u , W r , and W h s are trainable weight matrices; b z u , b r , and b h s are trainable bias vectors; and tanh · denotes the hyperbolic tangent function. The element-wise product operation ⊙ signifies the Hadamard product.
When multiple GRU layers are stacked, the hidden state of the previous layer is used as an input for the next layer, allowing for hierarchical learning of sequential patterns.
The structure of the network is shown in Figure 3. The maneuvering model estimation of a single moment mainly consists of the following two steps:
(Step 1)
Feature extraction: The input to the network consists of multiple observations of the position, including x o b s , y o b s , and z o b s , spanning from the past to the present [45]. At each time step k, the output includes observations of the velocity v o b s , k , yaw angle ψ o b s , k , and pitch angle θ o b s , k . While the position observations contain information about n x , n z , and ϕ , these characteristics are actually implicit in the second-order differentials between x o b s , y o b s , and z o b s . Moreover, due to measurement noise, using the observations directly as input for the GRU would mask actual changes in the states. To address this, we designed a “smoothing + first-order difference” feature extraction for the network. Specifically, we derive v o b s , k , ψ o b s , k , and θ o b s , k from the first-order differences of the filtered observations, which are then combined into a vector that serves as the input to the GRU layer. To reduce the impact of noise on estimating the motion model without accurately estimating the states of the centroid, we use the classical α β filtering for smoothing the observations, as expressed in Equations (36)–(39) [46].
Prediction:
x ^ α β , k | k 1 y ^ α β , k | k 1 z ^ α β , k | k 1 = x ^ α β , k 1 | k 1 y ^ α β , k 1 | k 1 z ^ α β , k 1 | k 1 + T S v ^ x , k 1 | k 1 v ^ y , k 1 | k 1 v ^ z , k 1 | k 1
v ^ x , k | k 1 v ^ y , k | k 1 v ^ z , k | k 1 = v ^ x , k 1 | k 1 v ^ y , k 1 | k 1 v ^ z , k 1 | k 1
Update:
x ^ α β , k | k y ^ α β , k | k z ^ α β , k | k = x ^ α β , k | k 1 y ^ α β , k | k 1 z ^ α β , k | k 1 + α x o b s , k x ^ α β , k | k 1 y o b s , k y ^ α β , k | k 1 z o b s , k z ^ α β , k | k 1
v ^ x , k | k v ^ y , k | k v ^ z , k | k = v ^ x , k | k 1 v ^ y , k | k 1 v ^ z , k | k 1 + β / T S x o b s , k x ^ α β , k | k 1 y o b s , k y ^ α β , k | k 1 z o b s , k z ^ α β , k | k 1
where ( · ^ ) α β , k | k 1 represents one-step prediction based on α β filtering, ( · ^ ) α β , k | k represents the smoothed state at time k, ( · ) o b s , k represents the observation at time k, and x , y , z and v x , v y , v z represent the position and velocity vector, respectively.
The first-order difference operation used in feature extraction can be derived as shown in Equation (40):
v o b s , k = x ^ α β , k | k , y ^ α β , k | k , z ^ α β , k | k T x ^ α β , k 1 | k 1 , y ^ α β , k 1 | k 1 , z ^ α β , k 1 | k 1 T 2 / T S θ o b s , k = arcsin z ^ α β , k | k z ^ α β , k 1 | k 1 [ x ^ α β , k | k , y ^ α β , k | k , z ^ α β , k | k ] T [ x ^ α β , k 1 | k 1 , y ^ α β , k 1 | k 1 , z ^ α β , k 1 | k 1 ] T 2 ψ o b s , k = arctan ( y ^ α β , k | k y ^ α β , k 1 | k 1 ) / ( x ^ α β , k | k x ^ α β , k 1 | k 1 )
In theory, the designed feature extraction can potentially learn the necessary features for maneuver detection. However, in practice, the expected increase in classification accuracy was not attained during model training. Further analysis of subsequent results revealed that the limited accuracy, which did not exceed 60%, predominantly stemmed from the narrow intervals between the feature samples and the relatively small absolute changes within individual data points. These factors posed challenges for the computer to precisely calculate the differences due to floating-point precision limitations. To tackle this issue, we experimented with amplifying the differences by multiplying each variable by 10 after calculating the first-order difference. This adjustment aimed to amplify the disparities between data points and facilitate accurate feature learning within a specified level of precision. Following this amplification, the model successfully converged during training, resulting in a classification accuracy of 99.73%.
(Step 2) Regarding model classification, the GRU layer receives a vector that comprises the velocity, yaw angle, and pitch angle. The maneuver model classification is then performed via a fully connected layer, a softmax layer, and a classification layer. The hidden layer contains 200 neurons, and each time step of the GRU layer generates an output.
During model training, the categorical cross-entropy loss function is applied as shown in Equation (41):
L = m = 1 M k = 1 K m n = 1 N μ m , k , n log ( μ ^ m , k , n )
where M represents the number of trajectories involved in the training process. Each trajectory has its own length and is represented by a sequence of K m data points. The variable N represents the number of categories, μ m , k , n represents the true probability of a given data point k at a trajectory m belonging to a category n, and μ ^ m , k , n indicates the predicted probability.
The model weights were initialized using Glorot initialization, and the ADAM optimizer was employed for optimization. A dropout rate of 0.5 was chosen for the experiment, with network connections being dropped between the final GRU output and the dense classifier. The network underwent 100 training epochs with a learning rate of 0.001 and a gradient threshold of 2.
In this paper, the training data utilized for model training were generated through simulation. The simulation process involved the following steps:
(1)
Target generation: A swarm of UAV targets, separated by distances ranging from 5 to 7 m, was generated with their geometric center flying along a predefined trajectory.
(2)
Maneuver and trajectory simulation: The trajectory of the center was divided into several time segments of varying lengths, with each segment randomly corresponding to a maneuver type. The trajectory of the center was simulated based on the selected maneuver type, and the corresponding motion model label was assigned.
(3)
Observation simulation: Observations of the swarm were simulated at a specific resolution, with an observation error having a standard deviation of 8 m along a single coordinate axis and an update frequency of 0.5 s. The centroid observation of the swarm was obtained by combining multiple measurements of the swarm using weighting.
The resulting training data consisted of observed centroid measurements with observation errors and corresponding motion model labels. The trajectory was divided into 50 segments, with a total duration of 75,812 s and an update interval of 0.5 s. The generated tracks are presented in Figure 4.

4.3. Estimation Fusion

At time step k, the UKF estimation and error covariance matrix of the ith model are denoted as x ^ k | k i and P k | k i , respectively. The weight assigned to model i by the maneuver estimation network is denoted as μ k i . The output state x ^ k | k and covariance P k | k of the multi-model filtering can be defined as shown in Equations (42) and (43), respectively:
x ^ k | k = i = 1 N μ k i x ^ k | k i
P k | k = i = 1 N μ k i P k | k i + [ x ^ k | k i x ^ k | k ] · [ x ^ k | k i x ^ k | k ] T
In [18], the weight μ k i is also determined by the considered deep learning network, which restricts its values to either one or zero. In practical scenarios, an alternative approach can be employed where the output of the softmax layer is utilized to assign weights to each model. The reason behind only considering binary values for μ is not explicitly provided [18]. Furthermore, the paper does not provide a direct comparison between the two methods. To address this gap, we will conduct a thorough analysis and comparison of both approaches in the Section 5. Our objective is to determine which method of assigning weights to models is more suitable for estimation fusion in the context of our study. By evaluating and contrasting these techniques, we aim to provide insights into the advantages and limitations of each approach, enabling us to make informed decisions regarding weight assignment in model fusion for estimation purposes.

5. Simulation

5.1. Simulation Settings

To verify the proposed algorithm, this study carries out simulation experiments that compare the filtering performance of maneuver model estimation and centroid tracking. In the comparative experiments for maneuver model estimation, we adopted the IMM algorithm and set the initial transition probability matrix as shown in Equation (44):
p i j = 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25
The initial mode probabilities were set as shown in Equation (45):
μ 1 : 2 : 3 : 4 = [ 0.25 0.25 0.25 0.25 ]
The validation trajectory’s length was 1213 s, with the observation update frequency set to 0.5 s. The σ R 2 , σ A 2 and σ E 2 values were set as 15 m, 0 . 1 , and 0 . 1 , respectively. We divided the validated trajectory into 30 segments, with a randomly generated maneuver mode used in each segment. The target’s velocity relative to the coordinate system’s origin fell within the range of (50, 120). The pitch variation for a single time segment was within the range π / 2 , π / 2 , while the yaw variation’s range was 0 , 2 π . We obtained the initial state values for filtering from Equation (23). The track generation process was similar to that in Section 4. Refer to Figure 5 for the tracking scenario. The simulation tool used in this paper is MATLAB 2021A.

5.2. Simulation Results and Analysis

Figure 6 illustrates the training process of the model estimation network. After a predetermined number of 100 iterations, the classification accuracy reached 99.5413%, and the loss function decreased to 0.02419. The network convergence performance was also better, as expected.
Figure 7 illustrates the difference in performance between the traditional IMM and proposed algorithms in estimating motion models. The GRU-assisted algorithm provided higher credibility and promptly reflected model transitions compared with the traditional IMM algorithm.
A 1000 run Monte Carlo simulation of position estimates is shown in Figure 8, which compares the root mean square error (RMSE) values of different algorithms. The IMM-EKF that employed the dynamical model proposed in this paper achieved more accurate state estimation than the Singer model. Overall, GRU-assisted filtering yielded a more accurate state estimation compared with the traditional IMM method. In Figure 9, we show the tracking results of the centroid, comparing the Singer-EKF, IMM-EKF, GRU-assisted filtering (1), and GRU-assisted filtering (2). GRU-assisted filtering (1) and GRU-assisted filtering (2) represent the model weight being zero or one and the model weight being the output of the softmax layer, respectively.
Furthermore, we observed that the reduction in error mainly occurred after model changes, demonstrating the ability of GRU-assisted motion model estimation to promptly reflect mode transitions. There was no significant difference between GRU-assisted filtering (1) and GRU-assisted filtering (2). The smaller RMSE of GRU-assisted filtering (2) may have been due to the retention of some states from other models during model transitions, which increased fault tolerance and mitigates the impact of model weight switching between 0 and 1 at incorrect time steps.

6. Experiments

6.1. Flight Experiments

To further evaluate the effectiveness of the proposed algorithm, flight experiments were conducted using three hexacopter UAVs developed in our laboratory [30]. As shown in Figure 10a, the UAVs were equipped with Langyu SUNNYSKY X5212S brushless motor multirotor motors, Tattu 30,000 mAh batteries, and the DJI A3 flight control system, providing flight endurance of 25 min. Three sets of flight experiments were carried out with different durations, namely 1194 s, 1243 s, and 1223 s. Each UAV was equipped with a BeiDou receiver (SinoGNSS M600mini) to record its real-time position. During each flight experiment, the three UAVs formed a swarm for cooperative flight, as shown in Figure 10b. The centroid trajectories of the swarm during cooperative flight, as depicted in Figure 11, consisted of multiple maneuvering segments. Within the maneuvering capabilities of the UAVs, the specific maneuvers performed were randomly generated by pre-programmed ground routines. Additionally, the YAOFENG RADAR YFR-01C was employed to record radar observations of the UAVs for predicting the motion model and trajectory filtering. The scan interval of the radar was 0.5 s.
The data processing for the experiment was divided into two parts: prediction of the maneuver models and trajectory filtering. In maneuver model prediction, two approaches were employed. The first involved directly using a network trained on a simulated dataset to validate the maneuver model with flight data, while the second approach utilized 70% of the flight data duration for training and the remaining 30% for validation. The training environment was set to mimic the simulation environment. Additionally, trajectory filtering was compared with the Singer-KF and IMM-EKF methods. Figure 12 illustrates the confusion matrix for the proposed deep learning approach in the flight experiment. The false alarm rate for mode estimation by the network trained in the simulation and the retrained network were found to be quite low, validating the algorithm’s capability to detect maneuver models. This also serves as preliminary evidence that the data generation process in the simulation closely approximated reality. Table 1 displays the root mean square error (RMSE) results for position and velocity estimation during the flight experiment. The results demonstrate that the algorithm can more accurately estimate the centroid trajectory of the UAV swarm. Furthermore, the proposed maneuver model and the GRU-assisted multi-model filtering significantly reduced tracking errors.
In practical scenarios, the high dynamic displacement between communicating parties can pose challenges to signal acquisition, synchronization, and stable transmission. Leveraging the spatial relationship between the parties’ displacements to assist with phase acquisition within communication signals can potentially lead to more stable or higher rates of data transmission under the same conditions. Enhanced accuracy in position estimation enables the trajectory information of high-dynamic UAV swarms’ centroids to serve as auxiliary data in communication systems. This auxiliary information, through specific mathematical transformations, can aid in rapid signal demodulation, precise compensation for Doppler frequency shifts, and tracking of communication beam directions within a phase-locked loop framework. Consequently, edge computing based on swarm UAVs can achieve more reliable or high-speed data interaction with the ground in high-dynamic environments.

6.2. Computational Complexity

In this section, the computational complexity of our 3DCDM-based GRU-MM algorithm is discussed in comparison with the Singer-KF and IMM-EKF algorithms by testing the average computation time for a single tracking iteration. To ensure a fair comparison, all algorithms were tested using the same Intel Core i7-8750H 2.2 GHz CPU and 16 GB of RAM. During the trajectory tracking process, our 3DCDM-based GRU-MM algorithm took an average of 17.4 ms per iteration, while the Singer-KF and IMM-EKF algorithms took 3.4 ms and 16.1 ms, respectively. Our 3DCDM-based GRU-MM algorithm achieved higher tracking accuracy without significantly sacrificing computational complexity. When the radar operated in the track-while-scan (TWS) mode with a scan interval of 0.5 s, our proposed method could be applied in real time.

7. Conclusions

Our proposed 3DCDM-based GRU-MM method aims to improve the tracking of maneuvering centroids in UAV swarms, which can significantly enhance the efficiency of mobile edge computing. By adopting a novel 3D coordinate-coupled dynamic model and multi-model filtering assisted by a GRU-based network, our method can better capture complex maneuvers and enhance state estimation accuracy. Furthermore, our proposed GRU-MM method achieved a high classification accuracy of 99.73%. The results demonstrate that our method outperformed traditional models or IMM algorithms in the state estimation of UAV swarms’ maneuvering centroids. The combination of our innovative tracking approach and mobile edge computing opens up new possibilities for advanced applications and services in various fields.
In the context of tracking the centroid of the UAV swarm, future research endeavors will involve exploring more intricate and precise maneuvering models to effectively capture the complex dynamic behaviors exhibited by UAV swarms in high-dynamic environments. This pursuit may involve introducing additional dimensions within dynamic models or considering the impact of uncertainty factors. Moreover, the integration of multiple sources of information will be a focal point, particularly in terms of effectively fusing data from various sensors and communication nodes to enhance the accuracy of UAV swarm state estimation.
Regarding the application of tracking results related to the centroids of UAV swarms, the subsequent step will involve exploring how trajectory information can be transformed to facilitate the enhancement of existing communication systems, specifically in the areas of phase-locked loop (PLL) operations, Doppler frequency estimation, beamforming techniques, and evaluating the implications of these advancements for edge computing.

Author Contributions

Conceptualization, X.L.; methodology, Y.C.; software, Y.C. and C.L.; validation, Y.C., M.W. and X.L.; formal analysis, X.L. and X.S.; investigation, Y.C., C.L. and X.S.; resources, M.W. and J.Z.; data calculation, M.W. and J.Z.; writing—original draft preparation, Y.C. and C.L.; writing—review and editing, M.W. and X.L.; visualization, M.W. and X.L.; supervision, X.L. and X.S.; project administration, M.W., J.Z. and X.S.; funding acquisition, M.W. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Science Foundation of China under grants 62001517, 61901502, and 62071352, the National Postdoctoral Program for Innovative Talents under grant BX20200101, the Electronic Information Equipment System Research National Defense Science and Technology Key Laboratory Fund under grant numbers 2023-HT-04 and 2023-HT-07, and the Research Foundation of the Key Laboratory of Spaceborne Information Intelligent Interpretation 2022-ZZKY-JJ-20-02.

Data Availability Statement

Data are contained within this article.

Conflicts of Interest

Authors Xiang Su was employed by the company China State Shipbuilding Corporation Limited. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. China Mobile. White Paper on 3D Coverage Network for Unmanned Aerial Vehicles Based on 5G Communication Technology; Technical Report; China Mobile: Xiamen, China, 2021. [Google Scholar]
  2. Xu, X.; Zhang, F.; Zhao, Y. Unmanned Aerial Vehicle Path-Planning Method Based on Improved P-RRT* Algorithm. Electronics 2023, 12, 4576. [Google Scholar] [CrossRef]
  3. McEnroe, P.; Wang, S.; Liyanage, M. A survey on the convergence of edge computing and AI for UAVs: Opportunities and challenges. IEEE Internet Things J. 2022, 9, 15435–15459. [Google Scholar] [CrossRef]
  4. Wu, W.; Zhou, F.; Wang, B.; Wu, Q.; Dong, C.; Hu, R.Q. Unmanned aerial vehicle swarm-enabled edge computing: Potentials, promising technologies, and challenges. IEEE Wirel. Commun. 2022, 29, 78–85. [Google Scholar] [CrossRef]
  5. You, J.; Ye, Z.; Gu, J.; Pu, J. UAV-Pose: A Dual Capture Network Algorithm for Low Altitude UAV Attitude Detection and Tracking. IEEE Access 2023, 11, 129144–129155. [Google Scholar] [CrossRef]
  6. Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Guo, G.; Ye, Q.; Jiao, J.; et al. Anti-uav: A large-scale benchmark for vision-based uav tracking. IEEE Trans. Multimed. 2021, 25, 486–500. [Google Scholar] [CrossRef]
  7. Zhou, L.; Leng, S.; Wang, Q.; Liu, Q. Integrated sensing and communication in UAV swarms for cooperative multiple targets tracking. IEEE Trans. Mobile Comput. 2023, 22, 6526–6542. [Google Scholar] [CrossRef]
  8. Chen, Y.; Jiao, Y.; Wu, M.; Ma, H.; Lu, Z. Group Target Tracking for Highly Maneuverable Unmanned Aerial Vehicles Swarms: A Perspective. Sensors 2023, 23, 4465. [Google Scholar] [CrossRef]
  9. Mihaylova, L.; Carmi, A.Y.; Septier, F.; Gning, A.; Pang, S.K.; Godsill, S. Overview of Bayesian sequential Monte Carlo methods for group and extended object tracking. Digit. Signal Process. 2014, 25, 1–16. [Google Scholar] [CrossRef]
  10. Li, X.R.; Jilkov, V.P. Survey of maneuvering target tracking. Part I. Dynamic models. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 1333–1364. [Google Scholar]
  11. Singer, R.A. Estimating optimal tracking filter performance for manned maneuvering targets. IEEE Trans. Aerosp. Electron. Syst. 1970, AES-6, 473–483. [Google Scholar] [CrossRef]
  12. Zhou, H.; Kumar, K. A’current’statistical model and adaptive algorithm for estimating maneuvering targets. J. Guid. Control Dyn. 1984, 7, 596–602. [Google Scholar] [CrossRef]
  13. Mehrotra, K.; Mahapatra, P.R. A jerk model for tracking highly maneuvering targets. IEEE Trans. Aerosp. Electron. Syst. 1997, 33, 1094–1105. [Google Scholar] [CrossRef]
  14. Semerdjiev, E.; Mihaylova, L.; Semerdjiev, T. Manoeuvring ship model identification and interacting multiple model tracking algorithm design. In Proceedings of the 1998 International Conference on Information Fusion, Las Vegas, NV, USA, 6–9 July 1998; pp. 968–973. [Google Scholar]
  15. Berg, R. Estimation and prediction for maneuvering target trajectories. IEEE Trans. Autom. Control 1983, 28, 294–304. [Google Scholar] [CrossRef]
  16. Bishop, R.H.; Antoulas, A. Nonlinear approach to aircraft tracking problem. J. Guid. Control Dyn. 1994, 17, 1124–1130. [Google Scholar] [CrossRef]
  17. Wang, C.; Yu, X.; Xu, L.; Wang, W. Energy efficient task scheduling based on traffic mapping in heterogeneous mobile edge computing: A green IoT perspective. IEEE Trans. Green Commun. Netw. 2023, 7, 972–982. [Google Scholar] [CrossRef]
  18. Wang, C.; Yu, X.; Xu, L.; Jiang, F.; Wang, W.; Cheng, X. QoS-aware offloading based on communication-computation resource coordination for 6G edge intelligence. China Commun. 2023, 20, 236–251. [Google Scholar] [CrossRef]
  19. Andrisani, D.; Kuhl, F.P.; Gleason, D. A nonlinear tracker using attitude measurements. IEEE Trans. Aerosp. Electron. Syst. 1986, AES-22, 533–539. [Google Scholar] [CrossRef]
  20. Li, X.R.; Jilkov, V.P. Survey of maneuvering target tracking. Part V. Multiple-model methods. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1255–1321. [Google Scholar]
  21. Lainiotis, D. Optimal adaptive estimation: Structure and parameter adaption. IEEE Trans. Autom. Control 1971, 16, 160–170. [Google Scholar] [CrossRef]
  22. Blom, H.A.; Bar-Shalom, Y. The interacting multiple model algorithm for systems with Markovian switching coefficients. IEEE Trans. Autom. Control 1988, 33, 780–783. [Google Scholar] [CrossRef]
  23. Bar-Shalom, Y.; Li, X.R. Estimation and Tracking-Principles, Techniques, and Software; Artech House, Inc.: Norwood, MA, USA, 1993. [Google Scholar]
  24. Li, X.R.; Bar-Shakm, Y. Mode-set adaptation in multiple-model estimators for hybrid systems. In Proceedings of the 1992 American Control Conference, Chicago, IL, USA, 24–26 June 1992; pp. 1794–1799. [Google Scholar]
  25. Li, X.R.; Bar-Shalom, Y. Multiple-model estimation with variable structure. IEEE Trans. Autom. Control 1996, 41, 478–493. [Google Scholar]
  26. Wang, C.; Deng, D.; Xu, L.; Wang, W. Resource scheduling based on deep reinforcement learning in UAV assisted emergency communication networks. IEEE Trans. Commun. 2022, 70, 3834–3848. [Google Scholar] [CrossRef]
  27. Li, X.R.; Zwi, X.; Zwang, Y. Multiple-model estimation with variable structure. III. Model-group switching algorithm. IEEE Trans. Aerosp. Electron. Syst. 1999, 35, 225–241. [Google Scholar]
  28. Li, X.R.; Zhang, Y.; Zhi, X. Multiple-model estimation with variable structure. IV. Design and evaluation of model-group switching algorithm. IEEE Trans. Aerosp. Electron. Syst. 1999, 35, 242–254. [Google Scholar]
  29. Li, X.R.; Zhang, Y. Multiple-model estimation with variable structure. V. Likely-model set algorithm. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 448–466. [Google Scholar]
  30. Moon, S.; Youn, W.; Bang, H. Novel deep-learning-aided multimodal target tracking. IEEE Sens. J. 2021, 21, 20730–20739. [Google Scholar] [CrossRef]
  31. Xie, G.; Sun, L.; Wen, T.; Hei, X.; Qian, F. Adaptive transition probability matrix-based parallel IMM algorithm. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 2980–2989. [Google Scholar] [CrossRef]
  32. Cosme, L.B.; D’Angelo, M.F.S.V.; Caminhas, W.M.; Camargos, M.O.; Palhares, R.M. An adaptive approach for estimation of transition probability matrix in the interacting multiple model filter. J. Intell. Fuzzy Syst. 2021, 41, 155–166. [Google Scholar] [CrossRef]
  33. Liu, J.; Wang, Z.; Xu, M. DeepMTT: A deep learning maneuvering target-tracking algorithm based on bidirectional LSTM network. Inf. Fusion 2020, 53, 289–304. [Google Scholar] [CrossRef]
  34. Kim, S.; Choi, J.; Kim, Y. Fault detection and diagnosis of aircraft actuators using fuzzy-tuning IMM filter. IEEE Trans. Aerosp. Electron. Syst. 2008, 44, 940–952. [Google Scholar]
  35. Lee, H.; Jung, J.; Choi, K.; Park, J.; Myung, H. Fuzzy-logic-assisted interacting multiple model (FLAIMM) for mobile robot localization. Rob. Auton. Syst. 2012, 60, 1592–1606. [Google Scholar] [CrossRef]
  36. Peng, H.; Zhang, X.; Li, H.; Xu, L.; Wang, X. An AI-Enhanced Strategy of Service Offloading for IoV in Mobile Edge Computing. Electronics 2023, 12, 2719. [Google Scholar] [CrossRef]
  37. Revach, G.; Shlezinger, N.; Ni, X.; Escoriza, A.L.; Van Sloun, R.J.; Eldar, Y.C. KalmanNet: Neural network aided Kalman filtering for partially known dynamics. IEEE Trans. Signal Process. 2022, 70, 1532–1547. [Google Scholar] [CrossRef]
  38. Chen, X.; Tao, Y.; Xu, W.; Yau, S.S.T. Recurrent neural networks are universal approximators with stochastic inputs. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 7992–8006. [Google Scholar] [CrossRef]
  39. Deng, L.; Li, D.; Li, R. Improved IMM algorithm based on RNNs. J. Phys. Conf. Ser. 2020, 1518, 012055. [Google Scholar] [CrossRef]
  40. Becker, S.; Hug, R.; Hübner, W.; Arens, M. An RNN-based IMM filter surrogate. In Proceedings of the Scandinavian Conference on Image Analysis, Norrköping, Sweden, 11–13 June 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 387–398. [Google Scholar]
  41. Fu, Q.; Lu, K.; Sun, C. Deep Learning Aided State Estimation for Guarded Semi-Markov Switching Systems with Soft Constraints. IEEE Trans. Signal Process. 2023, 71, 3100–3116. [Google Scholar] [CrossRef]
  42. Geng, W.d.; Wang, Y.q.; Dong, Z.h.; Geng, G.; Yang, F. Group-Target Tracking; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  43. Austin, F.; Carbone, G.; Falco, M.; Hinz, H.; Lewis, M. Automated maneuvering decisions for air-to-air combat. In Proceedings of the Guidance, Navigation and Control Conference, New Orleans, LA, USA, 11–13 August 1987; p. 2393. [Google Scholar]
  44. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  45. Zhang, Y.; Liang, Y.; Elazab, A.; Wang, Z.; Wang, C. Graph Attention Networks and Track Management for Multiple Object Tracking. Electronics 2023, 12, 4079. [Google Scholar] [CrossRef]
  46. Bar-Shalom, Y.; Li, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation: Theory Algorithms and Software; John Wiley & Sons: Hoboken, NJ, USA, 2002. [Google Scholar]
Figure 1. Scenario of UAV swarm-based edge computing.
Figure 1. Scenario of UAV swarm-based edge computing.
Electronics 13 01054 g001
Figure 2. 3D coordinate-coupled dynamic model of the UAV swarm’s centroid.
Figure 2. 3D coordinate-coupled dynamic model of the UAV swarm’s centroid.
Electronics 13 01054 g002
Figure 3. Network structure for maneuver model estimation.
Figure 3. Network structure for maneuver model estimation.
Electronics 13 01054 g003
Figure 4. Simulated UAV swarm trajectories for network training.
Figure 4. Simulated UAV swarm trajectories for network training.
Electronics 13 01054 g004
Figure 5. The tracking scenario used for validating the centroid trajectory tracking.
Figure 5. The tracking scenario used for validating the centroid trajectory tracking.
Electronics 13 01054 g005
Figure 6. Training process of the model estimation network.
Figure 6. Training process of the model estimation network.
Electronics 13 01054 g006
Figure 7. True mode and estimated mode probabilities of the maneuvering centroid. The topmost subgraph represents the true mode. The subgraphs below compare IMM estimates and GRU estimates for each model.
Figure 7. True mode and estimated mode probabilities of the maneuvering centroid. The topmost subgraph represents the true mode. The subgraphs below compare IMM estimates and GRU estimates for each model.
Electronics 13 01054 g007
Figure 8. RMSEs of the position estimates for Monte Carlo simulation of the proposed method and conventional method.
Figure 8. RMSEs of the position estimates for Monte Carlo simulation of the proposed method and conventional method.
Electronics 13 01054 g008
Figure 9. Comparison of different centroid tracking methods. The subgraph at the top left is a view of the entire flight path. The remaining three subgraphs are fragments selected from the track, and the corresponding areas are shown in the box in the figure.
Figure 9. Comparison of different centroid tracking methods. The subgraph at the top left is a view of the entire flight path. The remaining three subgraphs are fragments selected from the track, and the corresponding areas are shown in the box in the figure.
Electronics 13 01054 g009
Figure 10. (a) Picture of the UAV. (b) Picture of the cooperative flight.
Figure 10. (a) Picture of the UAV. (b) Picture of the cooperative flight.
Electronics 13 01054 g010
Figure 11. Flight trajectories with different durations of (a) 1194 s, (b) 1243 s, and (c) 1223 s.
Figure 11. Flight trajectories with different durations of (a) 1194 s, (b) 1243 s, and (c) 1223 s.
Electronics 13 01054 g011
Figure 12. Confusion matrix of three flight experiments with networks trained by simulation and actual observations. (a) The 1194 s flight with a simulated network. (b) The 1243 s flight with a simulated network. (c) The 1223 s flight with a simulated network. (d) The 1194 s flight with a retrained network. (e) The 1243 s flight with a retrained network. (f) The 1223 s flight with a retrained network.
Figure 12. Confusion matrix of three flight experiments with networks trained by simulation and actual observations. (a) The 1194 s flight with a simulated network. (b) The 1243 s flight with a simulated network. (c) The 1223 s flight with a simulated network. (d) The 1194 s flight with a retrained network. (e) The 1243 s flight with a retrained network. (f) The 1223 s flight with a retrained network.
Electronics 13 01054 g012
Table 1. The RMSEs of position and velocity estimation for the three flight experiments with GRU-assisted filtering, Singer-KF, and IMM-EKF.
Table 1. The RMSEs of position and velocity estimation for the three flight experiments with GRU-assisted filtering, Singer-KF, and IMM-EKF.
TrajectoryGRU-Assisted FilteringSinger-KFIMM-EKF
Flight1 position RMSE (m)2.52373.94193.2917
Flight1 velocity RMSE (m/s)1.68062.24512.0906
Flight2 position RMSE (m)2.70684.25433.3111
Flight2 velocity RMSE (m/s)1.88142.49852.1550
Flight3 position RMSE (m)2.69913.89353.5317
Flight3 velocity RMSE (m/s)1.69772.33912.1863
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Liu, X.; Li, C.; Zhu, J.; Wu, M.; Su, X. UAV Swarm Centroid Tracking for Edge Computing Applications Using GRU-Assisted Multi-Model Filtering. Electronics 2024, 13, 1054. https://doi.org/10.3390/electronics13061054

AMA Style

Chen Y, Liu X, Li C, Zhu J, Wu M, Su X. UAV Swarm Centroid Tracking for Edge Computing Applications Using GRU-Assisted Multi-Model Filtering. Electronics. 2024; 13(6):1054. https://doi.org/10.3390/electronics13061054

Chicago/Turabian Style

Chen, Yudi, Xiangyu Liu, Changqing Li, Jiao Zhu, Min Wu, and Xiang Su. 2024. "UAV Swarm Centroid Tracking for Edge Computing Applications Using GRU-Assisted Multi-Model Filtering" Electronics 13, no. 6: 1054. https://doi.org/10.3390/electronics13061054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop