Next Article in Journal
Full Solutions to Flow and Heat Transfer from Slip-Induced Microtube Shapes
Next Article in Special Issue
WPD-Enhanced Deep Graph Contrastive Learning Data Fusion for Fault Diagnosis of Rolling Bearing
Previous Article in Journal
Simulation and Experimental Study of a Piezoelectric Stack Energy Harvester for Railway Track Vibrations
Previous Article in Special Issue
Event-Based Optical Flow Estimation with Spatio-Temporal Backpropagation Trained Spiking Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SiamHAS: Siamese Tracker with Hierarchical Attention Strategy for Aerial Tracking

1
Changchun Institute of Optics, Fine Mechanics and Physics (CIOMP), Chinese Academy of Sciences, Changchun 130033, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Micromachines 2023, 14(4), 893; https://doi.org/10.3390/mi14040893
Submission received: 23 February 2023 / Revised: 26 March 2023 / Accepted: 30 March 2023 / Published: 21 April 2023
(This article belongs to the Special Issue Machine-Learning-Assisted Sensors)

Abstract

:
For the Siamese network-based trackers utilizing modern deep feature extraction networks without taking full advantage of the different levels of features, tracking drift is prone to occur in aerial scenarios, such as target occlusion, scale variation, and low-resolution target tracking. Additionally, the accuracy is low in challenging scenarios of visual tracking, which is due to the imperfect utilization of features. To improve the performance of the existing Siamese tracker in the above-mentioned challenging scenes, we propose a Siamese tracker based on Transformer multi-level feature enhancement with a hierarchical attention strategy. The saliency of the extracted features is enhanced by the process of Transformer Multi-level Enhancement; the application of the hierarchical attention strategy makes the tracker adaptively notice the target region information and improve the tracking performance in challenging aerial scenarios. Meanwhile, we conducted extensive experiments and qualitative or quantitative discussions on UVA123, UAV20L, and OTB100 datasets. Finally, the experimental results show that our SiamHAS performs favorably against several state-of-the-art trackers in these challenging scenarios.

1. Introduction

In the past decade, deep learning methods have been widely used and driven innovation in various research fields [1,2,3,4]. As an important task in the field of applications in deep learning, target tracking can be cross-applied with many disciplines, making it a hot topic of interest for researchers in many related fields, and is widely used in intelligent video surveillance [5], UAV applications [6], and assisted autonomous driving [7]. The task of traditional target tracking can be described as the following process: given only the initial location information and size of an arbitrary target in a video segment, the tracker needs to continuously estimate the subsequent motion state of the target without using other prior information. However, the actual video sequences often consist of scenarios such as scale variation, occlusion, and low resolution, which require high accuracy and robustness of the trackers, and it is a challenging task to ensure the performance of the tracker in complex scenes.
The existing target tracking methods could be divided into generative model-based tracking methods [8,9] and discriminative model-based tracking methods [10,11] according to the method of building target features. The former refers to the traditional methods which use the target’s feature and rely on manually designed features, such as the optical flow method [12], mean drift [13], etc., with generally poor tracking results. The latter refers to the discriminative model-based tracking methods that exploit both target motion foreground and background information, which have become mainstream over the last decade. The foreground indicates the pixels’ location where the target region in the feature maps is located, and the background refers to the pixel area occupied by the target motion background. Correctly distinguishing the foreground from the background is the key factor to locate the target object location and generate the predicted banding box. A step further, according to the implementation principle, it can be divided into correlation filter-based [14,15,16,17] and deep learning-based [18,19,20,21] tracking models. The key step in the correlation filter approach is to obtain the filter with optimal performance so that the search area and the target area could generate an accurate maximum response. For example, the KCF tracker [22], which is the cornerstone of the correlation filtering algorithms family, uses a similarity-matching strategy that transforms the image-matching task into a classification task of foreground and background and reduces the computational effort of the algorithm by using circular matrices and fast Fourier transforms. However, it cannot meet the requirements of tracking accuracy in fast background movement, small target object tracking, or other aerial scenarios. Among the features outputted by the deep learning-based feature extraction network, the shallow features contain location information, and in this regard, the shallower feature extraction network is more beneficial for determining the target position. The deeper features contain rich semantic information about the foreground and background in the images, which is useful for tracking in scenarios such as motion blur, low resolution, etc. The deep learning-based approach can automatically extract features at different depth levels through the feature extraction network and can effectively utilize the background information of the target motion, which makes it have strong feature extraction capability and robustness compared to traditional methods and correlation filter-based methods. In recent years, Siamese neural network-based trackers [23,24,25,26] have received extensive attention from researchers in the field.
Based on the above analysis, we propose a novel Siamese tracker based on the feature enhancement module of Transformer with the hierarchical attention strategy. The algorithm is mainly composed of the following four parts: feature extraction, feature fusion, similarity matching, and classification regression prediction head. The hierarchical attention strategy is the main characteristic of the feature fusion network. The self-attention mechanism can effectively improve the saliency of each branch, and the hierarchical attention strategy could achieve feature enhancement at different levels so that the features could include global contextual information and make it easier to utilize the features effectively. Experiments demonstrate that our SiamHAS can effectively enhance the tracking performance of Siamese neural networks in challenging scenarios.
The main contributions of our proposed approach can be summarized as follows:
  • We propose a target tracker with the strategy of hierarchical attention. After comparing with several state-of-the-art Siamese trackers, our SiamHAS reaps more stable robustness and precision in challenging scenes such as target occlusion, especially in various challenges in aerial scenarios, and achieves the desired goal.
  • In the feature fusion sub-network, a feature enhancement method integrating multiple attention strategies is introduced in this paper. With the channel context-aware mechanism, global context information can be included among different channels, which achieves feature enhancement in the channel dimension. The introduction of multiple modified Transformer encoders enhances the saliency of deeper features in each branch and effectively improves performance.
  • Extensive comparative experiments on our proposed SiamHAS are conducted on many benchmarks with complex and challenging scenarios. The experiments demonstrate that our algorithm obtains high robustness in challenging scenarios, and the hierarchical attention mechanism and Transformer module can strongly improve the performance of the Siamese network-based tracker.

2. Related Work

In this section, we briefly review the research related to our work in recent years, including a summary of the Siamese neural network trackers, attention mechanisms, and Transformers.

2.1. Siamese Trackers

The core idea of the Siamese trackers is to learn the information between the target template and the search region and then transform the tracking tasks into a similarity-matching problem between the two branches. The first Siamese tracker is SINT [27], and the subsequent series of Siamese trackers have a similar structure of using two branches of the same backbone network to generate feature maps. Then, after the mutual correlation operation, the best matching search region in the response maps is generated by the similarity matching network to complete the tracking, such as SiamFC [28], SiamDW [29], SiamDW-RPN [30], and SiamFC++ [31]. Inspired by Faster R-CNN [32] which proposed the RPN network to avoid the process of extracting feature maps at multiple scales, SiamRPN was proposed by Li et al. [33]. DaSiamRPN [34] achieves a degree of performance improvement but the depth of the extracted features is relatively shallow. SiamRPN++ [35] applies a deeper feature extraction network, but it uses a direct linear weighting fusion method without deeply exploring how to fully utilize the advantage of the deep network’s multiple layers feature. In recent years, more advanced methods such as HiFT [36], SiamBAN [37], SiamCAR [38], and Ocean [39] have been proposed one after another, all of which use deeper feature extraction networks. Although the above-mentioned trackers are applied to modern deep feature extraction networks, they do not take full advantage of the different levels of features, resulting in relatively poor performance in aerial scenarios, such as target occlusion, scale variation, and low-resolution target tracking.

2.2. Attention Mechanism

2.2.1. Introduction to Attention Mechanism

The attention mechanism [40,41,42] is one of the ways for humans to filter information efficiently. The main implementation of the attention mechanism can be described as follows: n input messages represented by a key-value pair can be written as [(k1,v1), (k2,v2)…, (kn,vn)]. The first step is the attention scoring function that uses the dot product model, which uses the inner product calculation to calculate the similarity of each covariate in the query object Q and K. In the second step, the attention scores are normalized to obtain the probability of each similarity, and the probability vector is called attention distribution. The normalization feature of the softmax function is used to highlight the weights of important elements and bring the network’s attention to the areas that need more attention. The overall implementation process of the attention mechanism can be summarized in the following equation. The attention score of each element in Q and K can be obtained as S. A is the final attention results, Q is the query variables, K is the key in the input key-value pair, and V is the value.
A ( Q , K , V ) = softmax ( S ( Q , K ) ) V
In the third step, the value is the weighted sum according to the coefficient of the attention distribution to aggregate information and obtain the final output result. The three-step process described above can be expressed as follows:
STEP 1 : S = s 1 , s 2 , , s n = Q K ; ( K = k 1 , k 2 , , k n ) STEP 2 : D = d 1 , d 2 , , d n = softmax ( s 1 , s 2 , , s n ) = e x p ( s i ) j = 1 n ( s j ) STEP 3 : A = i = 1 n d i v i
In this formula, the K represents the input keys in the key-value pair, in different conditions, it could be the original input values or input values after the linear transformation. S represents the attention score for the input key-value pair, si represents the attention score for each element in the input, and n represents the number of elements in the input. D represents the attention distribution for the input key-value pair, and it could be calculated by the softmax function with the input of S where the result is a normalized distribution of S. di represents the attention distribution for each element in the input. A represents the final attention results, it is the sum of all the products of vi which is the value element of all key-value pairs in the input and the corresponding attention distribution di.

2.2.2. Typical Attention Mechanisms in Computer Vision

Generally, self-attention [43] and non-self-attention are the typical classification method according to whether the ternary inputs are from the same source or not. In computer vision tasks, self-attention mechanisms are commonly used. According to the different dimensions of the extracted attention distribution, it can be generally classified into several forms of spatial [44], channel, and context [45] attention mechanisms. Taking the channel attention mechanism as an example, since the modern feature extraction network has deep layers, the change of channel dimension is essential, usually, the convolutional layers make significant contributions to the process, as shown in Figure 1.
The output feature map by each convolution kernel extracts a kind of feature of the target object, and each channel represents a matrix of each kind of feature. So, in the learning process, the parameters of the convolution kernel could adaptively vary and offer the net an appropriate structure. Similarly, the channel attention could finally obtain an attention distribution that can be recognized as a convolution kernel which could be optimized through the training process.

2.3. Transformer and Multi-Head Attention

Distinct from the attention mechanism in a broad sense, the self-attention mechanism stipulates a new restriction that the three elements of the input QKV must be homologous, which is usually known as the Transformer structure [46]. A typical Transformer structure definition had been released in “Attention is all you need” by Ashish Vaswani [47] in 2017. Its main implementation process can be summarized as the process of Scaled Dot-Product Attention, which is shown in Figure 2.
The mathematical expression of the model could be expressed as
Attention ( Q , K , V ) = softmax ( Q K T d k ) V Q u e r y : Q = W Q X K e y : K = W K X V a l u e : V = W V X
The matrix X represents the input data, and the values of Q, K, and V are derived from the same input X through matrix operations with three hyperparameters W Q , W K , W V that are first initialized and then learned through the process of optimization in training by applying an SGD optimizer with the strategy of random batch gradient descent. We use the dimension of the key, which is the number of its channels, as a parameter dk to scale the output of MatMul. This avoids the possibility of distortion in the softmax process which represents the process of obtaining a probability vector that stores attention distribution. The model of the multi-head attention mechanism describes a detailed introduction to the linear transformation matrix operations. The following Figure 3 and formulas illustrate the implementation process.
MultiHead ( Q , K , V ) = Concat ( head 1 , , head h ) W O headi = Attention ( Q W i Q , K W i K , V W i V )
In the above formula, Q, K, and V represent query, key, and value, respectively. For each head, headi is computed in the same process as the self-attention computation mentioned above, with each head initializing a different set of Wi parameter matrices. Furthermore, the results of these heads are finally concatenated together and then it is multiplied with a trainable parameter matrix W O to generate the final multi-head attention output. The multi-head attention calculates the similarity between Q and K, so the attention map could be obtained. Then, through the process of multiplying V with the normalized attention map, finally, we finish the process of enhancement for the inputs. To improve the various deficiencies of the existing methods, reference [48] adopts a Transformer-based feature fusion structure, which establishes the connection of global context, and the critical step is the powerful ability of the multi-head self-attention in the core module of the Transformer encoder.

3. Proposed Approach

In our proposed approach, we specifically introduce the proposed Siamese tracker and detail the overall structural composition of the algorithm and the principle of implementation. We focus on the structural composition of the hierarchical attention strategy for feature enhancement of the feature fusion sub-network then detail its basic modules and the important influence of the algorithm structure design. In particular, we focus on the implementation of the fusion strategy with channel, context, and multi-headed attention mechanisms, and the improvement of each part of the module, the architectural design, and implementation principles are presented in detail. Finally, we analyze the architectural design advantages of our proposed algorithm from the top-level perspective of the network.

3.1. The Overall Structure of SiamHAS

The proposed algorithm consists of the following four main modules: feature extraction backbone, feature fusion and enhancement network, similarity matching network, and classification regression prediction heads. The overall framework of our proposed tracker in this paper is shown in Figure 4.
In the Transformer Fusion Neck part, we propose the hierarchical attention strategy block and abbreviate it as the HAS block. It consists of the proposed channel context-aware attention module, which is abbreviated as the CCA module, and the Transformer multi-level enhancement module, which is abbreviated as the TME module, to realize the global correlation enhancement before feature fusion. Specifically, the last three-layer feature blocks output by the backbone are respectively used as the QKV ternary inputs of the three-level CCA modules, and the optimized three-layer feature blocks with the same dimension are the output. The following is a detailed description of the proposed CCA module.

3.1.1. The Proposed CCA Module: Channel Context-Aware Attention

For typical channel attention, if the dimension of the convolution output is C × H × W, the channel attention distribution learned through the network is stored in the C-dimensional vector, and the weight values of each layer are applied to all positions of each channel of the feature maps in turn. It could be the traditional way to realize the channel attention mechanism, but it still cannot make the most of the global context-aware information for modeling, and the efficacy is still comparatively low. Global context-aware information is a crucial factor for aerial tracking, as it could help to globally model using the correlation between consecutive frames to avoid tracking drift caused by temporary disappearance of the target in some frames or occlusion scenes. We propose a channel context-aware attention block applied to the fusion section, trying to apply the method of global contextual attention mechanism to different levels of features and realize the mining process of different depths of feature information from the dimension of channels. Specifically, we improve the optimization approach of attention feature maps in reference [45] from the perspective of theoretical analysis and experimental validation to balance the features at different levels. To adapt to the characteristics of the feature maps in aerial scenes and balance the effect of the information contained in different levels of feature maps, we concatenate the attention feature maps’ output with the initial input feature maps and adjust them with a 1 × 1 convolutional layer as the final output to achieve better feature utilization in aerial scenes and improve the performance of the tracker in challenging scenes. The channels containing the target features receive a larger attention scoring result through global contextual modeling to highlight the information of these channels and complete the pre-processing of the input data for the next process of Transformer multi-level enhancement. The specific implementation process of the proposed CCA module is displayed in Figure 5.
Specifically, the input feature maps are reshaped to obtain the new A in a different shape of C × N, which is then transposed to obtain AT and perform matrix multiplication of the two obtained feature maps. Then the attention map X in the shape of C × C is obtained after passing through the softmax layer. Through the process of matrix multiplication of XT and A in the shape of C×N, the attention scoring process can be applied to different channels based on different attention scores, and then reshape the results to the original shape to obtain attention feature maps O. The attention map involved in the calculation in this process could be expressed as follows:
X = softmax ( ( A Q ) T A K ) x K Q = e x p ( A Q A K ) i = 1 C e x p ( A Q A K )
where X represents the attention map, A Q and A K represent the features after the processing of the trainable parameter matrices W Q and W K , respectively. x K Q represents each element of the attention map, corresponding to the attention value of each location point. The output O and A are then fused through a process of channel cascade, and finally, the final output E is obtained through 1*1 convolution. The operation could be represented as
E = c o n v ( concat ( A , O ) )
O = conv ( A V · X )
where E represents the final output feature maps, A V represents the features after the processing of the trainable parameter matrix W V , O represents the attention feature maps after applying the attention map as a weight map to multiply with A V , conv represents the convolution operation, and concat represents the process of the concatenation. By the process of concatenation of the original input and the attention feature maps, the final output could contain multi-level information. This is an essential strategy to cope with aerial scenes such as occlusion or low resolution, because it is extremely demanding to fuse deep semantic information and shallow spatial information to complete the tracking process.
In summary, the proposed CCA module completes the optimization process of the source input data in the channel dimension by global modeling of the global context, using a self-attentive mechanism to obtain the channel correlations between any two different channels, and updating every channel graph using the strategy of the weighted sum of all channels.
After the tensor data is optimized by the CCA module, the feature information will be reduced by adjusting layers to avoid the large number of computational resources required in the subsequent TME module processing. Specifically, at the adjusting layers part, we take advantage of the multiple convolutional layers dimension adjustment, and the number of channels of the three-layer feature blocks after channel cascading is uniformly adjusted from the original [512, 1024, 2048] channels to [256, 256, 256] channels. This is to decrease the subsequent amount of parameters, and the dimensionality of the number of channels of the feature blocks will become 3 × 256. The following is a detailed description of the proposed TME module.

3.1.2. The Proposed TME Module: Transformer Multi-Level Enhancement

It is generally believed that the ability to effectively utilize the depth advantage of modern deep feature extraction networks is the key step to enhancing the existing Siamese trackers, and one of the most important steps is to effectively utilize different levels of deep features. Among the features extracted by the backbone at different depth levels, the shallow layer output features retain more concrete spatial information, such as edge, color, texture, etc., which helps to locate the target frame. The high-level feature maps after multiple convolutional and pooling layers contain deep semantic information, which can help the classification network to determine whether each pixel location is foreground or background. How to effectively utilize the feature information at different levels is a crucial process to enhance the model’s performance.
In aerial scenarios such as occlusion and motion blur, deep semantic feature information is detrimental to distinguishing the targets. In aerial tiny target detection and low-resolution scenarios, the pixel area of the target object has too few pixels to retain enough shallow spatial feature information, causing degradation in tracking performance. The existing methods often lack global interactions, and it is difficult to effectively exploit the association advantage of global contextual hierarchical features. To enhance the saliency of the hierarchical features at different levels to improve the performance of the tracker in the above aerial scenarios, we propose a TME module based on a hierarchical self-attentive strategy to cope with this problem.
Specifically, the pooling-based encoder structure is modified as the TME module in the second core attention in our proposed hierarchical attention strategy. As shown in Figure 6, the average pooling strategy is used in the Transformer encoder with the pooling strategy block, which is abbreviated as the TEP block, as a pretreatment strategy for refining the input data of the K and V parameters. We modify the typical Transformer encoder structure proposed in reference [48] as the basic architecture of the proposed TEP module through theoretical analysis and several experimental adjustments. We use two tandem TEP structures in the conv4 branch and a single TEP structure in the remaining two branches to achieve a balanced utilization of the features at different levels. Finally, we integrate the TME module consisting of four TEP modules into the structure design of the tracker to achieve better tracking performance. To further optimize to obtain a lighter Transformer-based structure for object tracking tasks, we replace the process of position coding of the typical Transformer structure with the zero-padding strategy using the location information of the inputs to ensure the integrity of input information.
The core process of the TEP block is processing the input ternary data, which is essentially the calculation process of the proposed pooling-based encoder, this could be mathematically expressed as
TEP ( Q , K , V ) = Norm ( I + MLP ( I ) ) I = Norm ( Q + MutiHead ( Q , Norm & Pooling ( K , V ) ) )
where QKV represent the ternary inputs, which are query variables, key-value variables, and value variables of the key-value pairs, respectively. MLP represents a fully connected feed-forward network, Pooling represents average pooling, and Norm represents the process of LayerNorm for smoothing the inputs. The following is a detailed description of how the proposed TME module is applied in our SiamHAS structure. Specifically, as shown in Figure 6, after the dimension adjustment, the feature block of 3,4,5 of each layer is flattened into sequence information using the convolution operation, and they are used as the inputs of the subsequent Transformer multi-level enhancement module, respectively, as the QKV ternary input of the three TEP blocks. The results of the TEP block whose input is conv4 are sent to another TEP block to achieve a balance of deep-level and shallow-level features through the calculation of double-layer self-attention. The three-branch TEP-based feature optimization calculation process constitutes the transformer multi-level enhancement process of our TME module. Taking the process of obtaining the results of the conv4 feature block F 4 as an example, it could be mathematically expressed as
R = TEP ( F 4 , F 4 , F 4 ) F 4 = TEP ( R , R , R )
where the intermediate variable R represents the features processed by the first TEP module and is used as the ternary input for the second TEP module.

3.1.3. The Similarity Matching Net and Downsize Feature Fusion

The output of the HAS block is the hierarchical fusion features that consist of shallow spatial information and deep semantic information. Finally, the number of channels and pixels of the feature maps will be further reduced by the downsizing feature fusion process after the similarity matching process, which greatly reduces the computational quantity in the following procedures of the feature processing In the similarity matching sub-network, the depthwise separable intercorrelation operation is used to calculate the response maps after each level of intercorrelation through a hierarchical convolution of the corresponding layers of 3,4,5 of the target template and the search region. The above process can be described as
R = φ ( X ) φ ( Z )
where represents the depthwise separable intercorrelation operation, and φ X and φ Z represent the hierarchical convolutional inputs corresponding to layers of 3,4,5 of the search region and template branch, respectively, which is also could be regarded as the hierarchical three-layer outputs of the TME blocks in the search region and template branch. The number of channels of the response map output is 256 × 3, which is still too large if it is directly used as the input of the subsequent classification and regression network. Therefore, we utilize multiple convolutional layers to decrease the dimensionality and adjust the dimension of the response maps. Specifically, the three blocks at different levels after the depthwise separable intercorrelation are sent to the downsizing block, which consists of multiple convolutional layers, to obtain 7 × 7 × 256 tensor for the template patch branch and 31 × 31 × 256 for the search region branch as the inputs of the following prediction heads. This operation decreases the number of model parameters and improves the speed of model derivation, ensuring a balance between model effectiveness and computational efficiency to the greatest extent. The above process can be represented by the following Figure 6.

3.1.4. The Prediction Heads

The prediction heads consist of three branches, including the CLS head, CEN head, and REG head shown in Figure 4, where the CLS head represents the classification network, the REG head represents the regression network, and the CEN head represents the centrality network. The role of the CLS head is to determine the attribute of every pixel position using the binary attributes of the target and background, to predict the attribute of the feature map location. The CEN head could obtain the prediction box’s centroid through multiple convolutional layers. The REG head could obtain the prediction box’s width and height to generate the tracking banding box through the continuous regression calculation of the network. Finally, the foreground and background are distinguished by the classification and regression network to obtain the target prediction frame. In the prediction head network, the response map generated after the intercorrelation operation can be regarded as a linear mapping of the input image in the space under the weight of the target template, and each position of the response map could be mapped to the corresponding position in the original image space of the search region, and the crucial process of mapping is the size of the receptive field of the linear transformation. Different from the algorithm based on region proposal network that uses fixed multi-scale frames for similarity matching to obtain multiple response maps, and then maps its maximum response area back to the corresponding search region position to achieve the prediction frame centroid and output the prediction frame by size regression, in this paper, we adopt the anchor-free mechanism to directly classify and regress the candidate frames at each position without setting a priori information to initialize the anchor frame size. Thus, we could break through the limitation of the scale and receptive field of fixed frames, minimize the limitation and interference of the set a priori information to the classification and regression network, avoid the introduction of multi-layer hyperparameter information which is difficult to optimally adjust, and ensure the computational efficiency of the classification and regression network.

3.1.5. The Loss Function

For all the heads introduced above, they receive an input of tensor in the shape of 25 × 25 × 256 after the process of similarity matching. For the final output, the classification head generates a classification feature map F c l s R H × W × 2 and the regression head generates a regression feature map F r e g R H × W × 4 , where W represents the width of the output feature map and H represents the height. In our case, they would be the same as the shape of the input tensor which is 25.
For the regression head, it outputs a feature map of size 25 × 25 with 4 channels, which records the distance from each corresponding location point to the 4 sides of the bounding box. This is noted as t ( i , j ) which is a 4D vector ( l , t , r , b ) , and could be calculated as follows:
t 0 ( i , j ) = l = x x 0 , t 1 ( i , j ) = t = y y 0 t 2 ( i , j ) = r = x 1 x , t 3 ( i , j ) = b = y 1 y
where ( x , y ) represents the coordinates of the location of the search region corresponding to the point ( i , j ) , and ( x 0 , y 0 ) and ( x 1 , y 1 ) represent the coordinates of the left-top and right-bottom corners of ground truth. The regression loss could be calculated by the following formula:
L r e g = 1 ( i , j ) I ( i , j ) ( i , j ) I ( i , j ) × I O U [ T ( i , j ) , t ( i , j ) ]
where T ( i , j ) represents the ground truth bounding box. IOU represents the common cross-merge ratio loss function, which is used to calculate the IOU loss between T ( i , j ) and t ( i , j ) , and I ( i , j ) is a judgment function, the value of it would be 1 when the sample point lies within the ground truth box, otherwise, it is 0.
For the classification head, it outputs a feature map of size 25 × 25 with 2 channels, which records the information of foreground and background at each point ( i , j ) in a 2D vector. The classification loss could be calculated using the method of cross-entropy loss:
L c l s = 0.5 × B C E L o s s ( δ p o s , I ) + 0.5 × B C E L o s s ( δ n e g , I )
where δ p o s and δ n e g represent the foreground and background scores stored in the 2D vector for each corresponding position in the search region. I represents the ground truth of the classification label at the location, and B C E L o s s represents the binary cross-entropy loss function.
For the centrality head, it outputs a feature map of size 25 × 25 with 1 channel called the centrality feature map F c e n R H × W × 1 , which records the centrality score C ( i , j ) of the corresponding position. The centrality score could be calculated by the formula:
C ( i , j ) = I × min ( l , t ) max ( l , t ) × min ( r , b ) max ( r , b )
where I is a judgment function, the value of it would be 1 when the sample point is in the foreground and lies within the ground truth box, otherwise, it is 0. The centrality loss is
L c e n = 1 ( i , j ) I I = 1 C ( i , j ) × log R ( i , j ) + ( 1 C ( i , j ) ) × log ( 1 R ( i , j ) )
where C ( i , j ) represents the predicted centrality score for a specific location and R ( i , j ) represents the actual centrality score for this position.
In summary, the overall loss function of the algorithm is as follows:
  L = L c l s + α 1 L c e n + α 2 L r e g
where L c l s , L c e n , and L r e g represent the classification loss, centrality loss, and regression loss, respectively. α 1 and α 2 are used as the weight hyperparameters to adjust the network and are set to 1 and 3, respectively, during the training process based on experimental and experiential evidence.

4. Experiments and Discussion

4.1. Experiments Settings

4.1.1. Experiments Environment and Parameter Settings

The experimental environment of the algorithms in the paper is set up as follows: the CUDA version is 11.8, and the Python 3.7 + pytorch 1.13 programming framework is used to train and verify the algorithm performance. The hardware platform configuration used is AMD Ryzen5 5600 @4.40 GHz processor for CPU and Nvidia GeForce RTX3080 for GPU, the total memory capacity is 16GB.
The parameters of the training process are initialized as follows: the tracker is trained by applying the SGD optimizer with random batch gradient descent, the batch size is set to 12, and the training result of the ResNet-50 network on ImageNet [49], a large computer vision image dataset, is used as the pre-trained weight model for the backbone network. The following strategy is used for learning rate setting: in the process, the learning rate modification strategy is set to weight decay, and the first 5 iterative processes are trained using a warm-up [50] strategy, where we initialize the learning rate at 0.001 and subsequently increase by 0.001 for each iteration. The model is trained with the strategy of a gradient descent of the learning rate after the warm-up. A total of 20 rounds of iterative processes are performed and the overall training time of the network is 65 h.

4.1.2. Datasets and Pretreatment

We train our SiamHAS network using the COCO [51], GOT-10K [52], and VID datasets as training sets with a sample size of 400,000 frames per round. To evaluate the performance in generalization and robustness of our algorithm from multiple perspectives, we use UAV123 [53] and OTB100 [54] datasets as test sets. A total of 100 video sequences containing various challenging scenes are available in the OTB100 dataset, and the UAV123 has 123 video sequences of aerial scenes, including multiple challenge categories such as scale transformation, occlusion, and viewpoint transformation in aerial scenes; it is the most frequently used test dataset for target tracking in aerial scenes.
We use the method of padding and cropping normalization to achieve scale-normalized preprocessing for images of different sizes in the datasets. The original images with a resolution less than 255 × 255 are filled with edge pixels to be the desired input resolution and used as the search region branch input of the network. Meanwhile, the ground truth information that is read from the Jason file is used as a basis to cut out the 127 × 127 area that contains most of the target’s information in the original picture, and then it is used as the template branch input. The purpose is to normalize the input and improve the training efficiency. Since the network structure will transform the input image into a smaller resolution 256 × 256 feature map, the normalized dataset will undoubtedly retain more target feature information and enhance the training speed of the algorithm and the effectiveness of the model. The following Figure 7 shows the detailed process of pre-processing described in the previous section.

4.2. Evaluation Metrics

The frequently used evaluation method for target tracking which is called one-pass evaluation (OPE) is generally used to initialize the first frame with the information in the ground truth. Then, the tracker could obtain the average accuracy and precision in the following tracking process. The common evaluation parameters for target tracking mainly include center position error and overlapping area ratio. The details will be described in the following subsections. Here Figure 8 shows a visual interpretation of the two kinds of evaluation metrics introduced above.

4.2.1. Precision

Precision is a common evaluation metric for single target tracking tasks. Precision is defined as the percentage of the frames where the Euclidean distance between the centroid of the target location generated by the tracker and the centroid of the manually labeled target is less than a set threshold. Usually, the threshold is set to 20 pixels, and the Euclidean distance within 20 pixels is regarded as the criterion for successful tracking. Specifically, the Euclidean distance calculation method between them is
ρ = ( x 2 x 1 ) 2 + ( y 2 y 1 ) 2
Different thresholds are chosen as the criteria for judging the accuracy requirements, so different percentages of the frames that meet the thresholds can be obtained, and a precision curve can be drawn.

4.2.2. Success Rate

The success rate is determined by calculating the intersection ratio of the pixels in the region of the prediction box and the ground truth box, as shown in Figure 9 below. This is the ratio of the intersection area of the red box and the green box in the figure below to the total area surrounded by the two boxes.
Specifically, the overlap score OS is calculated as:
O S = | x y | | x y |
The bounding box generated by the tracker is recorded as x, the ground truth box is recorded as y, and |·| indicates the number of pixels in the region. When the OS of a frame is above the set threshold, it could be regarded as successful tracking, and the success rate is calculated by the percentage of the total successful frames to all the frames. The value range of OS is 0~1, so a success rate evaluation curve could be drawn.

4.3. Performance Comparison

4.3.1. UAV123 Benchmark

The UAV123 benchmark is composed of 123 image sequences filmed by an aerial vehicle. It consists of various challenging characteristics, including tiny targets, low resolution, occlusion, and so on. Success and precision plots are the two kinds of evaluation methods for the trackers. For precision plots, the threshold is set to less than 20 pixels to judge if the specific frame meets the requirement. Precision refers to the percentage of frames whose CLE meets the requirements in the total frames. The formulas are as follows.
C L E = ( x p r x g t ) 2 + ( y p r y g t ) 2 f = { 1 , C L E < 20 0 , C L E 20 P r e c i s i o n = i = 1 N f N
In the formulas, (xpr, ypr) and (xgt, ygt) refer to the center position of the prediction box and ground truth, respectively. For success rate, the overlap ratio S threshold is set to larger than 0.5 to judge if the specific frame meets the requirement and can be regarded as successful tracking. The success rate is calculated by the percentage of the number of frames whose overlap ratio S meets the requirements in the total frames. The overlap represents the IOU of the prediction box and the ground truth of each frame. The final result takes the average value of all the frames. The formulas are as follows.
S = I O U | R p r R g t | | R p r R g t | f = { 1 , I O U 0.5 0 , I O U < 0.5 P r e c i s i o n = i = 1 N f N
The success plots and precision plots of the one-pass evaluation on the trackers are shown in Table 1. Our proposed tracker SiamHAS is observed to obtain a success rate of 0.627 and a precision of 0.820. Compared with other Siamese series trackers such as SiamBAN, which also has a similar overall structure design, our SiamHAS obtains a large improvement of 2.3% and 2.5% in success rate and precision.

4.3.2. Attribute-Based Discussion on UAV123

To illustrate the performance improvement of the proposed method under challenging scenarios such as low resolution, occlusion, and scale variation, we conduct an attribute-based experiment on UAV123. The results of the attribute-based evaluation on UAV123 are shown in Figure 10, demonstrating that the proposed SiamHAS achieves the best evaluation results in all the attribute-based challenges. For success rate, the proposed SiamHAS obtains the best results, including full occlusion (0.416), low resolution (0.489), partial occlusion (0.553), and scale variation (0.612). In the standardized process of the one-pass evaluation for target tracking algorithms, for the success plots, the threshold of overlap ratio is set to 50 percent to judge whether a particular frame is tracked successfully or not, and the percentage of the frames judged to be successfully tracked to the total number of frames is defined as the success rate, which is recorded in the box on the right side corresponding to the different trackers, and the color ranking is constant and represents the success rate ranking of different trackers. Similarly, for the precision plots, the threshold of the location error is set to 20 pixels to judge whether a particular frame meets the requirements of expected precision or not. For the precision plot, the proposed SiamHAS also obtains the best performance among all the comparison attribute challenges mentioned above. Compared with the other Siamese series trackers such as SiamCAR, the proposed SiamHAS obtains significant performance improvement in full occlusion and low resolution challenges with 1.9% and 2.5% in precision, and 2.2% and 2.3% in success rate. Furthermore, it is worth noting that further attribute-based experiments show that the proposed method seems to show a degradation in the performance of background clutter scenarios. This might be caused by the fact that the proposed hierarchical attention mechanism does not balance well the effect of background noise while improving the saliency of the target. To sum up, the attribute-based discussion demonstrates that the proposed SiamHAS displays significant improvement in the performance of challenging aerial scenes and how to overcome the effect of background clutter will be a key consideration in our future work.

4.3.3. Case Visual Comparison on UAV123

To intuitively show the improvement of our proposed tracker SiamHAS in coping with challenging scenes such as low resolution, occlusion, and scale variation, we choose six cases from the UAV123 dataset, that consist of the above issues. Results are then compared to other Siamese series trackers and several state-of-the-art trackers, and the case visual comparison is shown in Figure 11.

4.3.4. UAV20L Benchmark

UAV20L is a dataset for aerial image and video analysis, which contains 20 long-time sequences of UAV images in different urban scenes. These image sequences cover a wide range of scenes, including city streets, parks, buildings, and natural landscapes, and are widely used as a benchmark to test the tracker’s long-time tracking performance. To further verify the robustness of the proposed SiamHAS in aerial scenarios, especially the performance when coping with long-time target tracking, we conducted experiments on the UAV20L benchmark to more comprehensively illustrate the performance improvement of the designed tracker in coping with aerial scenarios, and to intuitively represent the performance improvement of the tracker in optimizing the design for aerial scenarios. The success rate and precision of the one-pass evaluation on the trackers are shown in Table 2. Our proposed tracker SiamHAS is observed to obtain a success rate of 0.573 and a precision of 0.745. Compared with the other state-of-the-art Siamese series trackers such as SiamAPN++ [55], our SiamHAS obtains an improvement of 1.3% and 0.9% in success rate and precision. From the experimental results, we could draw the conclusion that the proposed SiamHAS tracker outperforms other state-of-the-art trackers in aerial long-term target tracking scenarios. Additionally, we have comprehensively evaluated the performance of the proposed tracker in aerial scenarios after the UAV20L benchmark experiment, which strongly supports the effectiveness of the innovative design and the improvement methods of the proposed tracker in adapting to the characteristics of aerial scenarios.

4.4. Quantitative Analysis

To validate the generalization ability of our SiamHAS in non-aerial scenarios for other types of challenges, we choose four typical scenarios in the OTB100 dataset including (a) fast motion, (b) motion blur, (c) occlusion, and (d) out of view for the attribute-based comparison experiments with some other state-of-the-art algorithms. Figure 12 displays the attribute-based evaluation and demonstrates that the proposed SiamHAS achieves the best performance compared with the other trackers that participate in the comparison on the attributes-based challenges. SiamHAS greatly improves the performance, especially in the two key scenes of occlusion and out of view.

4.5. Discussion and Ablation Studies

4.5.1. Case Discussion on OTB100

To further validate the generalization capability of our tracker SiamHAS and intuitively demonstrate the performance improvement in coping with challenging scenarios such as occlusion, motion blur, and out of view, we choose four cases from the OTB100 dataset. The results are compared to several state-of-the-art trackers, such as SiamCAR, SiamRPN++, SiamRPN, and SiamFC, the visual comparison pictures are shown in Figure 13.

4.5.2. Heat Map Experiment

To more intuitively show the comparison between the other state-of-the-art Siamese series trackers’ and our SiamHAS’ focusing region and show the performance improvement of our tracker in coping with complex scenes such as similar object interference and motion blur, we perform a comparison-based heat map analysis with the recent and related approach SiamCAR under equal conditions. Specifically, we choose five cases in the OTB100 dataset, and the areas that draw the attention of the trackers are shown in Figure 14.

4.5.3. Modules Ablation Experiment

To provide a deeper insight into how each of the proposed modules contributes to the overall network performance improvement, we carry out an ablation study on multiple benchmarks. The intention of this is to show how our proposed modules contribute to the performance of the tracker, and how much performance improvement every module could bring to the tracker. As shown in Table 3, we use the symbol √ to indicate the specific block has been added to the tracker architecture; × represents the opposite.

5. Conclusions

In our present article, we propose a novel Siamese tracker SiamHAS to improve the shortcomings of tracking drift and tracking failure in challenging aerial scenarios such as occlusion, aspect ratio change, low resolution, and scale variation. The proposed HAS module adopts a hierarchical attention strategy, which can take advantage of semantic and spatial features to break through the limitations of the receptive field of traditional convolutional neural networks and achieve multi-layer feature enhancement. In the channel attention part, we achieve global feature enhancement in the channel dimension, and through the channel context-aware process, we make the model adaptively distribute the attention scores of each channel based on global modeling. Therefore, there is an information exchange between different channels, which makes each channel contain more global contextual information and achieves a more effective feature optimization. In addition, self-attention enhancement of deep information is achieved by multiple parallel multi-headed attention networks, which enhances the model’s capability to distinguish foreground and background while balancing deep and shallow feature information. Additionally, we have performed many analyses and experiments on UAV123, UAV20L, and OTB100 datasets, experiments based on attribute challenges, case discussion, visualization, ablation experiments, etc. The experiments show that SiamHAS obtains significant capability improvement in all challenges compared with several state-of-the-art Siamese trackers; it also shows some generalization ability in the dataset of non-aerial scenes. Compared with the Siamese series tracker SiamCAR, it shows significant performance improvement for aerial scenarios, which also verifies that the proposed method can effectively cope with challenging scenarios such as occlusion, aspect ratio change, low resolution, etc., without losing much performance in real-time while ensuring the accuracy and robustness of the algorithm. In the case of using a deeper feature extraction network, the algorithm in this paper maintains a frame rate above 45 FPS, which meets the requirement of real-time tracking.

Author Contributions

Conceptualization, F.L. and Q.C.; methodology, F.L.; project administration, J.L.; resources, X.W. and C.L.; software, F.L.; validation, F.L.; investigation, F.L., Q.C., and J.L.; visualization, F.L.; writing—original draft preparation, F.L.; writing—review and editing, Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 61905240.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are grateful for the anonymous reviewers’ critical comments and constructive suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zuo, C.; Qian, J.; Feng, S.; Yin, W.; Li, Y.; Fan, P.; Han, J.; Qian, K. Deep learning in optical metrology: A review. Light Sci. Appl. 2022, 11, 39. [Google Scholar] [CrossRef] [PubMed]
  2. Li, J.; Jiang, S.; Song, L.; Peng, P.; Mu, F.; Li, H.; Jiang, P.; Xu, T. Automated optical inspection of FAST’s reflector surface using drones and computer vision. Light Adv. Manuf. 2023, 4, 1. [Google Scholar] [CrossRef]
  3. Huang, L.; Luo, R.; Liu, X.; Hao, X. Spectral imaging with deep learning. Light Sci. Appl. 2022, 11, 61. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, Y.; Liu, T.; Singh, M.; Çetintaş, E.; Luo, Y.; Rivenson, Y.; Larin, K.V.; Ozcan, A. Neural network-based image reconstruction in swept-source optical coherence tomography using undersampled spectral data. Light Sci. Appl. 2021, 10, 155. [Google Scholar] [CrossRef]
  5. Kim, J.S.; Yeom, D.H.; Joo, Y.H. Fast and robust algorithm of tracking multiple moving objects for intelligent video surveillance systems. IEEE Trans. Consum. Electron. 2011, 57, 1165–1170. [Google Scholar] [CrossRef]
  6. Wang, S.; Jiang, F.; Zhang, B.; Ma, R.; Hao, Q. Development of UAV-based target tracking and recognition systems. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3409–3422. [Google Scholar] [CrossRef]
  7. Cao, J.; Song, C.; Song, S.; Xiao, F.; Zhang, X.; Liu, Z.; Ang, M.H., Jr. Robust Object Tracking Algorithm for Autonomous Vehicles in Complex Scenes. Remote Sens. 2021, 13, 3234. [Google Scholar] [CrossRef]
  8. Wu, H.; Zhou, J. Privacy Leakage of SIFT Features via Deep Generative Model Based Image Reconstruction. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2973–2985. [Google Scholar] [CrossRef]
  9. Nakada, S.; Meguro, T.; Zhao, Q. An Improved Approach for Generative Model-Based Product Image Anomaly Detection. In Proceedings of the 2022 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Yeosu, Republic of Korea, 26–28 October 2022; pp. 1–4. [Google Scholar]
  10. Lei, Y.; Ding, X.; Wang, S. Visual Tracker Using Sequential Bayesian Learning: Discriminative, Generative, and Hybrid. IEEE Trans. Syst. Man Cybern. Part B 2008, 38, 1578–1591. [Google Scholar]
  11. Jiang, N.; Su, H.; Liu, W.; Wu, Y. Discriminative Metric Preservation for Tracking Low-Resolution Targets. IEEE Trans. Image Process. 2012, 21, 1284–1297. [Google Scholar] [CrossRef]
  12. Shi, J.; Tomasi. Good features to track. In Proceedings of the 1994 IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
  13. Comaniciu, D.; Ramesh, V.; Meer, P. Real-time tracking of non-rigid objects using mean shift. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No. PR00662), Hilton Head, SC, USA, 15 June 2000; Volume 2, pp. 142–149. [Google Scholar]
  14. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual Object Tracking Using Adaptive Correlation Filters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
  15. Du, F.; Liu, P.; Zhao, W.; Tang, X. Correlation-guided attention for corner detection based visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2020, Seattle, WA, USA, 13–19 June 2020; pp. 6836–6845. [Google Scholar]
  16. Nishimura, H.; Nagai, Y.; Tasaka, K.; Yanagihara, H. Object tracking by branched correlation filters and particle filter. In Proceedings of the 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR), Nanjing, China, 26–29 November 2017. [Google Scholar]
  17. Su, Y.; Liu, J.; Xu, F.; Zhang, X.; Zuo, Y. A Novel Anti-Drift Visual Object Tracking Algorithm Based on Sparse Response and Adaptive Spatial-Temporal Context-Aware. Remote Sens. 2021, 13, 4672. [Google Scholar] [CrossRef]
  18. Nam, H.; Han, B. Learning Multi-Domain Convolutional Neural Networks for Visual Tracking. In Proceedings of the 29th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 4293–4302. [Google Scholar]
  19. Zhu, K.; Zhang, X.; Chen, G.; Tan, X.; Liao, P.; Wu, H.; Cui, X.; Zuo, Y.; Lv, Z. Single object tracking in satellite videos: Deep Siamese network incorporating an interframe difference centroid inertia motion model. Remote Sens. 2021, 13, 1298. [Google Scholar] [CrossRef]
  20. Huang, B.; Xu, T.; Shen, Z.; Jiang, S.; Zhao, B.; Bian, Z. SiamATL: Online Update of Siamese Tracking Network via Attentional Transfer Learning. IEEE Trans. Cybern. 2021, 52, 7527–7540. [Google Scholar] [CrossRef] [PubMed]
  21. Deng, A.; Liu, J.; Chen, Q.; Wang, X.; Zuo, Y. Visual Tracking with FPN Based on Transformer and Response Map Enhancement. Appl. Sci. 2022, 12, 6551. [Google Scholar] [CrossRef]
  22. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef]
  23. He, A.; Luo, C.; Tian, X.; Zeng, W. Towards a better match in siamese network based visual object tracker. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
  24. An, N.; Yan, W.Q. Multitarget tracking using Siamese neural networks. ACM Trans. Multimid. Comput. Commun. Appl. 2021, 17, 75. [Google Scholar] [CrossRef]
  25. Cheng, S.; Zhong, B.; Li, G.; Liu, X.; Tang, Z.; Li, X.; Wang, J. Learning to filter: Siamese relation network for robust tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 21–24 June 2021; pp. 4421–4431. [Google Scholar]
  26. Guo, Q.; Feng, W.; Zhou, C.; Huang, R.; Wan, L.; Wang, S. Learning dynamic siamese network for visual object tracking. In Proceedings of the IEEE International Conference on Computer Vision 2017, Venice, Italy, 22–29 October 2017; pp. 1763–1771. [Google Scholar]
  27. Tao, R.; Gavves, E.; Smeulders, A.W.M. Siamese instance search for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1420–1429. [Google Scholar]
  28. Wang, L.; Ouyang, W.; Wang, X.; Lu, H. Visual tracking with fully convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision 2015, Santiago, Chile, 11–18 December 2015; pp. 3119–3127. [Google Scholar]
  29. Zhang, Z.; Peng, H. Deeper and wider siamese networks for realtime visual tracking. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4591–4600. [Google Scholar]
  30. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  31. Xu, Y.; Wang, Z.; Li, Z.; Yuan, Y.; Yu, G. Siamfc++: Towards robust and accurate visual tracking with target estimation guidelines. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12549–12556. [Google Scholar]
  32. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  33. Fan, H.; Ling, H. Siamese cascaded region proposal networks for real-time visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–21 June 2019; pp. 7952–7961. [Google Scholar]
  34. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Distractor-aware siamese networks for visual object tracking. In Proceedings of the European Conference on Computer Vision (ECCV) 2018, Munich, Germany, 8–14 September 2018; pp. 101–117. [Google Scholar]
  35. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J.S. Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 16–20. [Google Scholar]
  36. Cao, Z.; Fu, C.; Ye, J.; Li, B.; Li, Y. HiFT: Hierarchical Feature Transformer for Aerial Tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15457–15466. [Google Scholar]
  37. Chen, Z.; Zhong, B.; Li, G.; Zhang, S.; Ji, R. Siamese box adaptive network for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2020, Seattle, WA, USA, 16–18 June 2020; pp. 6668–6677. [Google Scholar]
  38. Guo, D.; Wang, J.; Cui, Y.; Wang, Z.; Chen, S. SiamCAR: Siamese fully convolutional classification and regression for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2020, Seattle, WA, USA, 16–18 June 2020; pp. 6269–6277. [Google Scholar]
  39. Zhang, Z.; Peng, H.; Fu, J.; Li, B.; Hu, W. Ocean: Object-aware anchor-free tracking. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 771–787. [Google Scholar]
  40. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  41. Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.-H.; Martin, R.R.; Cheng, M.-M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  42. Brauwers, G.; Frasincar, F. A general survey on attention mechanisms in deep learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 3279–3298. [Google Scholar] [CrossRef]
  43. Yang, B.; Wang, L.; Wong, D.; Chao, L.S.; Tu, Z. Convolutional self-attention networks. arXiv 2019, arXiv:1904.03107. [Google Scholar]
  44. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV) 2018, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  45. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  46. Zhang, D.; Zhang, H.; Tang, J.; Wang, M.; Hua, X.; Sun, Q. Feature pyramid transformer. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 323–339. [Google Scholar]
  47. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  48. Xing, D.; Evangeliou, N.; Tsoukalas, A.; Tzes, A. Siamese transformer pyramid networks for real-time UAV tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision 2022, Waikoloa, HI, USA, 3–8 January 2022; pp. 2139–2148. [Google Scholar]
  49. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  52. Huang, L.; Zhao, X.; Huang, K. Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1562–1577. [Google Scholar] [CrossRef] [PubMed]
  53. Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for uav tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 445–461. [Google Scholar]
  54. Wu, Y.; Lim, J.; Yang, M.H. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
  55. Cao, Z.; Fu, C.; Ye, J.; Li, B.; Li, Y. SiamAPN++: Siamese attentional aggregation network for real-time UAV tracking. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 3086–3092. [Google Scholar]
Figure 1. A multi-channel convolution process where a single convolutional kernel convolves with the corresponding feature maps, and the individual layer pixel positions are added to produce an output single-channel feature map.
Figure 1. A multi-channel convolution process where a single convolutional kernel convolves with the corresponding feature maps, and the individual layer pixel positions are added to produce an output single-channel feature map.
Micromachines 14 00893 g001
Figure 2. A typical self-attention mechanism implementation process where MatMul represents the dot product operation of the input matrix Q and K.
Figure 2. A typical self-attention mechanism implementation process where MatMul represents the dot product operation of the input matrix Q and K.
Micromachines 14 00893 g002
Figure 3. The implementation process of multi-head attention computing.
Figure 3. The implementation process of multi-head attention computing.
Micromachines 14 00893 g003
Figure 4. The overall framework of the proposed SiamHAS tracker.
Figure 4. The overall framework of the proposed SiamHAS tracker.
Micromachines 14 00893 g004
Figure 5. The structure design of the proposed CCA module.
Figure 5. The structure design of the proposed CCA module.
Micromachines 14 00893 g005
Figure 6. The schematic of the process of Transformer multi-level enhancement, depthwise separable intercorrelation, and downsize operation.
Figure 6. The schematic of the process of Transformer multi-level enhancement, depthwise separable intercorrelation, and downsize operation.
Micromachines 14 00893 g006
Figure 7. The process of datasets pretreatment: padding and cropping normalization. The (a) original images are adjusted through the padding process to be 255 × 255 as the inputs of the (b) search region branch X, and then are cropped to 127 × 127 as the inputs of the (c) template patch Z branch using the ground truth information.
Figure 7. The process of datasets pretreatment: padding and cropping normalization. The (a) original images are adjusted through the padding process to be 255 × 255 as the inputs of the (b) search region branch X, and then are cropped to 127 × 127 as the inputs of the (c) template patch Z branch using the ground truth information.
Micromachines 14 00893 g007
Figure 8. Visual presentation of (a) center position error and (b) overlapping area ratio. The centroid error between the target template and the search area prediction output and the overlapping area ratio of the two areas are used as the core evaluation calculation indicators.
Figure 8. Visual presentation of (a) center position error and (b) overlapping area ratio. The centroid error between the target template and the search area prediction output and the overlapping area ratio of the two areas are used as the core evaluation calculation indicators.
Micromachines 14 00893 g008
Figure 9. Visual presentation of the process of computing overlapping area ratio.
Figure 9. Visual presentation of the process of computing overlapping area ratio.
Micromachines 14 00893 g009
Figure 10. Attribute-based experiment results corresponding to 5 scenes on UAV123, including (a) full occlusion; (b) low resolution; (c) partial occlusion; (d) scaled variation; (e) background clutter.
Figure 10. Attribute-based experiment results corresponding to 5 scenes on UAV123, including (a) full occlusion; (b) low resolution; (c) partial occlusion; (d) scaled variation; (e) background clutter.
Micromachines 14 00893 g010aMicromachines 14 00893 g010b
Figure 11. Case study of the proposed SiamHAS and several state-of-the-art trackers on the UAV123 dataset. From the top to bottom, the pictures show the visual case experiment results on the image sequences of (a) Bird1, (b) Car9, (c) Car14, (d) Group1, (e) Person11, and (f) Uav1. Red—Ours, Green—SiamCAR, Blue—SiamRPN++, Black—SiamRPN, Pink—SiamFC.
Figure 11. Case study of the proposed SiamHAS and several state-of-the-art trackers on the UAV123 dataset. From the top to bottom, the pictures show the visual case experiment results on the image sequences of (a) Bird1, (b) Car9, (c) Car14, (d) Group1, (e) Person11, and (f) Uav1. Red—Ours, Green—SiamCAR, Blue—SiamRPN++, Black—SiamRPN, Pink—SiamFC.
Micromachines 14 00893 g011aMicromachines 14 00893 g011b
Figure 12. Experimental results corresponding to 4 typical scenarios on OTB100, including (a) fast motion; (b) motion blur; (c) occlusion; (d) out of view.
Figure 12. Experimental results corresponding to 4 typical scenarios on OTB100, including (a) fast motion; (b) motion blur; (c) occlusion; (d) out of view.
Micromachines 14 00893 g012aMicromachines 14 00893 g012b
Figure 13. Case study on OTB100 for our SiamHAS and several state-of-the-art trackers. From the top to bottom, the pictures show the visual case experiment results on the image sequences of (a) Bird1, (b) lemming, (c) Coupon, and (d) human4. Red—Ours, Green—SiamCAR, Blue—SiamRPN++, Black—SiamRPN, and Pink—SiamFC.
Figure 13. Case study on OTB100 for our SiamHAS and several state-of-the-art trackers. From the top to bottom, the pictures show the visual case experiment results on the image sequences of (a) Bird1, (b) lemming, (c) Coupon, and (d) human4. Red—Ours, Green—SiamCAR, Blue—SiamRPN++, Black—SiamRPN, and Pink—SiamFC.
Micromachines 14 00893 g013aMicromachines 14 00893 g013b
Figure 14. The heat map experiment visual results correspond to 5 scenes, including (a) blot subset; (b) box subset; (c) cardark subset; (d) crowds subset; (e) dog subset.
Figure 14. The heat map experiment visual results correspond to 5 scenes, including (a) blot subset; (b) box subset; (c) cardark subset; (d) crowds subset; (e) dog subset.
Micromachines 14 00893 g014aMicromachines 14 00893 g014b
Table 1. UAV123 benchmark comparison table, the table shows the success rate and precision results of 9 trackers including ours. The proposed SiamHAS performs favorably against state-of-the-art trackers. The best method is shown in red, and the second is shown in green.
Table 1. UAV123 benchmark comparison table, the table shows the success rate and precision results of 9 trackers including ours. The proposed SiamHAS performs favorably against state-of-the-art trackers. The best method is shown in red, and the second is shown in green.
TrackerOursCGACDSiamCARSiamRPN++SiamBANHiFTSiamDWSiamRPNSiamFC
Success0.6270.6200.6150.6110.6040.5890.5820.5360.478
Precision0.8200.8150.8040.8040.7950.7870.7760.7720.697
Table 2. UAV20L benchmark comparison table, the table shows the success rate and precision results of 8 trackers including ours. The proposed SiamHAS performs favorably against state-of-the-art trackers. The best method is shown in red, and the second is shown in green.
Table 2. UAV20L benchmark comparison table, the table shows the success rate and precision results of 8 trackers including ours. The proposed SiamHAS performs favorably against state-of-the-art trackers. The best method is shown in red, and the second is shown in green.
TrackerOursSiamAPN++HiFTSiamAPNSiamFC++SiamRPNDaSiamRPNSiamFC
Success0.5730.5600.5530.5390.5330.5280.5470.423
Precision0.7450.7360.7360.7210.6950.6960.6760.629
Table 3. Comprehensive evaluation results with or without the proposed blocks on multiple datasets. Success-U and PRE-U represent the success rate and precision on the UAV123 benchmark; Success-O and PRE-O represent the success rate and precision on the OTB100 benchmark. FPS represents the selected results on the GOT-10K benchmark. The best method is shown in red, and the second is shown in green.
Table 3. Comprehensive evaluation results with or without the proposed blocks on multiple datasets. Success-U and PRE-U represent the success rate and precision on the UAV123 benchmark; Success-O and PRE-O represent the success rate and precision on the OTB100 benchmark. FPS represents the selected results on the GOT-10K benchmark. The best method is shown in red, and the second is shown in green.
NO.CCATMESuccess-UPRE-USuccess-OPRE-OFPS
1××0.5990.7730.6760.87951
2×0.6010.7890.6810.88149
3×0.6150.8040.6860.88747
40.6270.8200.6930.89645
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, F.; Liu, J.; Chen, Q.; Wang, X.; Liu, C. SiamHAS: Siamese Tracker with Hierarchical Attention Strategy for Aerial Tracking. Micromachines 2023, 14, 893. https://doi.org/10.3390/mi14040893

AMA Style

Liu F, Liu J, Chen Q, Wang X, Liu C. SiamHAS: Siamese Tracker with Hierarchical Attention Strategy for Aerial Tracking. Micromachines. 2023; 14(4):893. https://doi.org/10.3390/mi14040893

Chicago/Turabian Style

Liu, Faxue, Jinghong Liu, Qiqi Chen, Xuan Wang, and Chenglong Liu. 2023. "SiamHAS: Siamese Tracker with Hierarchical Attention Strategy for Aerial Tracking" Micromachines 14, no. 4: 893. https://doi.org/10.3390/mi14040893

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop