Next Article in Journal
Quantum Advantages of Teleportation and Dense Coding Protocols in an Open System
Next Article in Special Issue
Attention and Pixel Matching in RGB-T Object Tracking
Previous Article in Journal
Robust Semi-Quantum Summation over a Collective-Dephasing Noise Channel
Previous Article in Special Issue
A High-Precision Two-Stage Legal Judgment Summarization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PACR: Pixel Attention in Classification and Regression for Visual Object Tracking

School of Information Engineering, Wuhan University of Technology, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(6), 1406; https://doi.org/10.3390/math11061406
Submission received: 17 February 2023 / Revised: 10 March 2023 / Accepted: 11 March 2023 / Published: 14 March 2023

Abstract

:
Anchor-free-based trackers have achieved remarkable performance in single visual object tracking in recent years. Most anchor-free trackers consider the rectangular fields close to the target center as the positive sample used in the training phase, while they always use the maximum of the corresponding map to determine the location of the target in the tracking phase. Thus, this will make the tracker inconsistent between the training and tracking phase. To solve this problem, we propose a pixel-attention module (PAM), which ensures the consistency of the training and tracking phase through a self-attention module. Moreover, we put forward a new refined branch named Acc branch to inherit the benefit of the PAM. The score of Acc branch can tune the classification and the regression of the tracking target more precisely. We conduct extensive experiments on challenging benchmarks such as VOT2020, UAV123, DTB70, OTB100, and a large-scale benchmark LaSOT. Compared with other anchor-free trackers, our tracker gains excellent performance in small-scale datasets. In UAV benchmarks such as UAV123 and DTB70, the precision of our tracker increases 4.3% and 1.8%, respectively, compared with the SOTA in anchor-free trackers.
MSC:
68T01; 68T07; 68T45

1. Introduction

Single Object Tracking is a fundamental work in computer vision with so many applications in various fields such as autonomous driving, human–computer interactions, video annotation [1,2,3], etc. Although brilliant methods have been proposed during the past decade, there are still some difficulties in practical applications due to challenging factors, such as fast motions, background clutter, occlusions, deformation, and other difficulties, which urges us to develop trackers with strong adaptiveness and robustness.
In recent years, no matter if they are transformer-based or Siamese-based, or correlation-based, most trackers which adopt the standard tracking-by-detection framework are divided into two subtasks: feature extraction and object location. The feature extraction subtask obtains the feature which helps to distinguish the differences and similarities between the template frame and the subsequent search frame. While the object location subtask will give the location of the tracking object with the help of the result computed by the previous feature’s extraction subtask. Usually, the computing method is correlation in the Discriminative Correlation Filters (DCFs) [4,5,6,7,8,9] or convolution in the Siamese trackers [10,11,12,13,14,15]. Recently, transformer-based trackers [16,17,18,19,20] such as OSTrack [21] adopt the dot product as the computing method. Those three computing methods are utilized to calculate the above similarities.
In the tracking phase, the tracker could get the target’s approximate location by the maximum value of the calculated similarities. But in the training phase, most trackers only can use the proximity principle, which means they have to regard the rectangle area near the target as the ground truth shown in Figure 1, especially the anchor-free trackers. It can be found that the location rules are inconsistent between the training and tracking phase.
The Pixel Attention Classification and Regression (PACR) tracker is proposed to solve the inconsistencies. With the above-mentioned tracking methods, our tracker has remarkable performance in some challenging scenarios. Our contribution can be summarized in three aspects:
  • A new attention is proposed, named pixel attention, to tackle the incongruence between the training and tracking phases, which differs from the channel and spatial attention. The pixel attention makes the tracker focus on the object which is most like the target to adjust the ground truth and reduce the training time.
  • The Acc branch is put forward, which will further expand the advantages of pixel attention during the tracking phase to improve the tracking result in the object location subtask. The score of the Acc which is trained by the predicting acceleration can rectify the tracker’s outputs more accurately.
  • A novel end-to-end tracker is proposed with anchor-free baseline. Using the joint classification and regression to compute the loss score in the training phase, the score of the Acc branch can be better integrated into the Cls (classification) branch and the Reg (regression) branch. As a result, the quality of the tracker can be improved notably.

2. Related Work

This section will briefly review some related work about single visual object tracking. Mainly, some strategies are introduced, such as anchor-free mechanisms and attention in object tracking.

2.1. Discriminant Correlation Filter

DCF-based tracker learning a target model to locate a target by minimizing an objective has attracted much attention in object tracking because of its finite computation cost and robustness. MOSSE [22] (Minimum Output Sum of Squared Error filter) is the first to introduce the correlation algorithm into object tracking. Then CSK [11] and the Kernel Correlation Filter [5] (KCF), which introduced the ridge regression and approximate dense sampling method based on cyclic shifts into the tracker, made DCF-based trackers more popular. SRDCF [6], STRCF [7], and ARCF [8] introduce the spatial penalty and temporal penalty to solve the boundary effect and improve the robustness and accuracy of the tracker Auto-track [9], and this [23] makes the UAV-based object track become the practical application stage from the research stage; the latter also summarizes the problems existing in the current DCF-based UAV trackers.

2.2. Siamese-Based Tracker

Siamese-based trackers have attracted attention recently and have become very popular in this field. The first one proposed to solve the tracking problem is the SiamFC [10], tracking the target by learning the similarity between the template and search region. Inspired by SiamFC, many works have followed and put forward updated models. The correlation filter layer and online tracking are introduced into the Siamese-based tracker to improve the accuracy in CFNet [24]. DaSiam [13] successfully explored the effect of sample selection on the tracker, which improves the robustness of the tracker. With the development of RPN networks, SA-Siam [11] built a two-fold Siamese network with a semantic branch and an appearance branch. They are trained separately to keep the heterogeneity of features but tested in a combination way. SA-Siam has proved that the training phase can differ from the testing phase. SiamPRN++ [14], using ResNet-50 as the backbone, has declared that feature extraction may affect the tracking result. Meanwhile, it randomly shifts the object location in the search region during training to eliminate the center bias, surprisingly improving tracking accuracy.

2.3. Anchor-Free Mechanism

Although anchor-based trackers have many benefits, such as boosting the tracking result, improving the tracking accuracy, or making the IOU score higher, some difficulties of anchor-based mechanisms still need to be solved. For example, the performance of most anchor-based trackers is susceptible to the hyper-parameter, and the size and the aspect ratio of the bounding box of these trackers are fixed. Moreover, the design of anchors is also related to the object category. In recent years, the anchor-free mechanism has gained more and more attention because of its simplicity and efficiency. The anchor-free mechanism doesn’t have trouble like the anchor-based mechanism in anchor design. It was first put forward in DenseBox to solve the object detection, then CornerNet [25] and ExtremeNet [26] further studied the anchor-free mechanism, expanding the predicted corner to 2 or 4 points. FCOS [27] directly indicates the bounding box with the help of a new strategy named center-ness. Ocean [28], SiamFC++ [15], and SiamCAR [29] are the anchor-free trackers, calculating the loss by IoU-loss in the training phase, which increase the accuracy of the predicted bounding box.

2.4. Attention in Object Tracking

Nowadays, attention has gained more and more attraction in single visual object tracking and achieved impressive results. Most attention-based trackers apply attention mechanisms in two ways. Guided by the principles of CNN, the attention mechanism can be a backbone to extract features. For example, OSTrack [21] directly uses ViT [30] as the backbone to extract features; Swin [31], PVT [32], PVTv2 [33], and MixFormer [17] also adopt the same strategy. They prove an attention-based backbone can gain the same benefit as a traditional CNN-based backbone. However, the CNN-based backbone performs better if it takes the similar architecture and strategies of a recent study named ConvNeXt [34]. Merely fusing the features extracted from the backbone is the other method. TCTrack [18] delivers the information from the last frame to the current search frame with the help of attention. STMTrack [16] directly applies a self-attention encoder to the computation of the search-frame feature by combining the historic search frames. AiATrack [20], ToMP [35], and Transformer Tracking [36] are also this way. Although they can track targets robustly without online learning, padding design is difficult to initialize for direct use, such as Stark, which will incur enormous computational costs. Considering this, we propose a new method to combine the attention mechanism and improve the tracker’s accuracy.

2.5. Guidance

In recent work of anchor-free Siamese trackers, most trackers select the rectangle region close to the center of the target as a positive target during training. It is not a good way to set positives for deformed objects because the tracker will study more parameters from the background information, which affects the final tracking result. Pixel attention is proposed to solve this problem by keeping the training phase consistent with the tracking phase. In order to inherit the benefit of the pixel attention, we build a new Acc branch in the classification and regression head.

3. Method

This section will exhaustively introduce our PACR tracker in three aspects: the basic architecture, pixel-attention mechanism, and Acc branch. The whole network is shown in Figure 2, containing a Siamese network based on the CNN for feature extraction and a multi-head tracking network for object location. The prior network is similar to other Siamese networks, such as SiamCAR and SiamOA [37]. Meanwhile, a novel training method is adopted in the latter network.

3.1. Basic Architecture

The PACR divides the tracking task into two subtasks, similar to the classical Siamese-based tracker SiamCAR. One is extracting the feature from the tracking target. Another is obtaining the final target location with the help of the feature extracted from the first network.
As shown in Figure 2, the feature extracted network has two similar branches with the same parameter. The front is the search branch, and the bottom is the template branch. The PACR adopts the feature from layer 2, layer 3, and layer 4 of resnet50. The feature score A w × h × c c will be obtained after the inverse convolution of the feature map, which is calculated by combining the feature extracted from the prior network. We adopt a novel design idea in the following classification prediction network. During the training and tracking phase, the PACR adopts two similar but not identical networks to ensure that our network has a high training speed and can also ensure that the performance of tracking is not significantly affected; the Acc branch was uniquely proposed to improve the regression results.
In the training phase, unlike the original design that directly uses the score A w × h × c c related to the search frame and the template frame, we will input the related calculation score A w × h × c p of the previous frame and the correlation score A w × h × c c obtained from the current search frame and the template into the PAM (pixel-attention module), generate a related feature score that is more focused on tracking the target, input it into the Acc branch, and input it into the other three branches to get four scores ( a c c i , j , c l s i , j , c e n i , j , l o c i , j ) , and then the total loss will be calculated with a joint classification–regression alternating refinement strategy combining several branches for optimization.
We get the coarse location generated by the above network:
p o s c o a r s e = arg m a x i , j { ( 1 λ d ) a c c i , j × c l s i , j × c e n i , j × p i , j + λ d H i , j }
H is the cosine window, λ d is a super-parameter, and the combination of these can suppress large changes in the bounding box and position. p is a penalty updated with the search frame changed. After the coarse position is obtained, the accurate position p o s r e f i n e = ( x r e f i n e , y r e f i n e ) will be calculated in a coarse-to-fine refine strategy [29]. l o c i , j is a vector of w × h × 4 with each subvector representing the distance from the corresponding position to the four sides of the bounding box. So the height and width of the predicting box can be performed as follows:
w = l o c ( x r e f i n e , y r e f i n e , 0 ) + l o c ( x r e f i n e , y r e f i n e , 2 ) 2 , h = l o c ( x r e f i n e , y r e f i n e , 1 ) + l o c ( x r e f i n e , y r e f i n e , 3 ) 2
The final predicting box generated in the tracking phase is
B p r e d = ( x r e f i n e w 2 , y r e f i n e h 2 , w , h )
The whole tracking process is shown in Figure 3.

3.2. Pixel Attention

The current method of determining the positive samples in the training phase of the anchor-free frame tracker is to directly learn all the features within a certain range of the ground truth bounding box, as shown in Figure 1. The sample from this method is unreasonable. Directly confirming the positive sample according to the distance from the center point will make the tracker learn more background information during tracking. Although this method generalizes the model of the tracking net, in the case of tracking targets with obvious changes in characteristics such as human movement, the reg score and cls score will be disturbed by the background due to sufficient learning of the background region, which will make the tracking result worse. More importantly, when the anchor-free tracker generates the bounding box in the tracking phase, it uses the maximum position of the correlation operation between the search frame and the template frame, and finally determines the boundary according to the reg score of the maximum position. Taking the method mentioned above in training will lead to inconsistency between the tracking and training phases because the position of the maximum response point is considered a positive sample during tracking for the anchor-free tracker.
Above all, the tracker should determine the more accurate target area as positive samples such as the deep color area in Figure 1. We use a simple encoder named pixel attention, similar to a self-attention method in STMT. However, different from STMT, pixel attention uses the self-attention module to obtain the weight of each pixel so that the tracker can be trained. When we can pay more attention to the target, instead of fitting part of the background, the idea of pixel attention is closer to PSA [38], using self-attention to obtain the attention of each pixel. The PACR focuses on improving the precision of positive samples, helping enhance the consistency of the tracker in the training phase and the tracking phase. The pixel-attention module (PAM) is shown in Figure 4.
The main function of pixel attention is to make the tracker fit the position that is very likely to be the target point during the training phase and suppress the score of the non-target point area. The formula of the PAM is
w i , j = exp [ ( f i . q f . j k ) / s ] k ( f k . q f . j k ) / s , f i = ( f v ) i T w
The feature map f q , f k , f v is generated from three different Conv layers, and w i , j is the weights of the corresponding points, where i is the index of each pixel on f q R T H W × C , and j is the index of each pixel on f q R C × H W . Here, s is a scaling factor to prevent the exp function from overflowing numerically. Following [39], we set s to C , where C is the feature dimensionality of f .

3.3. ACC Branch

The PAM has two functions: one is to get a better positive sample into the trainer for training; another is to strengthen the feature and improve the ability to express the feature. To fully exploit the advantages of the PAM module in the tracking phase, we propose the Acc branch, which can optimize the scores generated by other branches through the joint classification–regression alternating refinement method from SiamOA [37]. In this branch, our tracker adopts the strategy of this paper [40]. The ideal result of the Acc branch is shown below:
A c ( i , j ) = Φ ( P i , j 3 + P i , j 1 2 × P i , j 2 ( P i , j 3 + P i , j 1 2 × P i , j 2 ) a v g )
P i represents the distance of the point of the same position from the bounding box in frame i, and for the moving trend is significantly greater than the movement trend of other positive sample averages. It can be considered that this point is most likely the background. Otherwise, it is always considered as the tracking target. Φ ( x ) is performed as
Φ x = x 0 x 1 1 x x 1 0 o t h e r .
The Acc branch learns the movement trend of the target, and by combining acc score and cen score, the final feature map can be closer to the real situation. The new version of the PAM is shown in Figure 5.
The input is changed from only the correlation score f c of the current search frame and the template frame to the correlation score f p of the previous frame and the template frame and the correlation score f c of the current frame and the template frame. The reason for our design is that the moving distance of the target is limited between two frames, so using f p f c to obtain w will more accurately describe the tracked target. The loss of this branch is shown below:
L a c c = 1 N A c ( i , j ) × log A w × h × 1 a c c ( i , j ) + ( 1 A c ( i , j ) ) × log ( 1 A w × h × 1 a c c ( i , j ) )
By combining the score of Acc and the score of Cen, PACR can track the target occluded by other objects, as shown in Figure 6.

4. Experiment and Results

4.1. Tracking Detail

Our network is implemented in Python with PyTorch and trained on 4 RTX3090 GPUs. The proposed PACR applies the ResNet50 as the feature extraction backbone with the parameter pre-trained on ImageNet for the classification task. During the training phase, we adopt the PAM with the batch size set as 256 and epochs set as 50. For fair comparison [14,29], the parameter of the backbone network is frozen in the first 20 epochs. For the later 30 epochs, the last three blocks of ResNet50 are unfrozen for training. Specifically, our training datasets include COCO [41], imageNet VID [42], LaSOT [43], ImageNet DET [42], and YouTube-BB [44] for experiments on UAV123 [45], DTB70 [46], VOT2020 [47], and OTB100 [48]. The whole training phase takes about 56 h. For the large-scale experiments of LaSOT, we train our network in another 20 epochs with the training dataset LaSOT. We froze the parameters unfrozen in the first ten epochs and unfroze them in the later ten epochs. In the testing phase, following the idea we mentioned above, the PAM does not need to be taken part. The super-parameter which can be found in our code is just coarse-tuned. For evaluating the tracking results, we take advantage of the measurements from their official website, which may differ from each other.

4.2. Results on VOT2020

The VOT dataset is an authoritative dataset in single visual object tracking. The VOT2020 [47] dataset contains 60 public sequences with different challenging factors. VOT2020 has two different challenges: VOT-ST2020 and VOT-RT2020. The former is the VOT short-term tracking challenge under appearance variation, occlusion, and clutter. The latter is VOT short-term real-time challenge VOT-RT2020 under time constraints. The main metrics to evaluate the tracker are EAO (Expected Average Overlap), A (Accuracy), R (Robustness), and AUC (Average Unsupervised Accuracy). We compare our tracker with some typical and SOTA trackers from previous years. The results of our tracker are shown in Table 1
Compared with SiamCAR, our model baseline, we improve by 1.1% and 1.7%, respectively, in accuracy and AUC. Additionally, from Table 1, our tracker is better in accuracy, which indicates our PAM can help the tracker get more target information in the training phase, helping the tracker track better in the testing phase.

4.3. Results on OTB100

OTB100, proposed in 2015, is a classical benchmark in single visual object tracking containing 100 sequences with substantial variations collected and annotated from the commonly used tracking sequence. It has different attributes labeled with video sequences, such as deformation, motion blur, out-of-view, background clutter, etc. The metric of OTB100 is the precision plot and success plot. The former indicates the distance between the center of the bounding box and the center of the ground truth, and the latter means the IOU scores of the ground-truth boxes and the bounding box.
We compared our tracker with current mainstream methods, including anchor-free based trackers such as SiamCAR, SiamBan, Ocean, and TCTrack. The result has shown that our tracker is better than other anchor-free-based trackers. More information about the comparison is shown in Figure 7

4.4. Results on UAV123

UAV123 contains 123 sequences extracted from 91 videos and more than 110K frames. The sequences from this dataset are fully annotated. The same as OTB100, it also has many challenges for tracking, such as illumination variation, occlusions, and fast motion. The best difficult challenge in this dataset is the scale variation. We compare it with some state-of-the-art trackers in the UAV123 benchmark, and as with the OTB100 evaluation, we compare it with other anchor-free trackers.
The result of the UAV123 dataset evaluation is shown in Figure 8. Our tracker outperforms all other anchor-free-based trackers in precision, as the figure shows. Compared with CAR, our tracker improves the score by 1.8% in precision plots of OPE.

4.5. Results on DTB70

DTB70, similar to UAV123, which is composed of 70 challenging UAV image sequences, primarily proposes the problem of severe UAV motion. The DTB70 dataset has tracking problems, such as various cluttered scenes and objects of different sizes. We compare our tracker with some SOTA trackers in this dataset, such as TCTrack and AutoTrack. The results come from their paper.
As shown in Figure 9, compared with TCTrack, the state-of-the-art UAV tracker in 2022, our tracker improves by 1.4% in success and decreases by 0.3% in precision. PACR has 55 FPS in this dataset which is much faster than TCTrack. Compared with CAR, our tracker obtains a 4.3 percentage point increase and one percentage point increase, respectively, in success and precision.

4.6. Results on LaSOT

LaSOT is a large-scale single object tracking dataset with high-quality annotations. The dataset contains over 3.52 million manually annotated frames and 70 classes, each with 20 tracking sequences. Its training set consists of 1400 long videos, and its testing set consists of 280 long videos, with an average of 2500 frames per video. LaSOT can test the robustness of trackers properly, which have many difficulties tracking challenges such as occlusions, out-of-view, etc. The PACR is tested on the testing set. As shown in Table 2, compared with other trackers whose data are from their papers, PACR has shown better performance to these trackers. Our tracker also gets 65 FPS on average, which is much better than other trackers.

4.7. Complexity Analysis

Compared with other trackers, such as SiamRPN++, our model has fewer parameters. The total FLOPS of SiamRPN++ is 59.56G, which is more than our model. The complexity of each part of our model is shown in the Table 3.
In order to reduce the model complexity, some small tricks [51] were used during the training and tracking phases. The first trick is that data enhancement performed on the images to ensure that the model’s generalization ability is not affected by some adjustments to the resolution of the input images. Secondly, transfer learning is used to reduce the model training time with improved training efficiency and generalization ability of the model. Third, the total number of parameters has been reduced using Siamese structures at the backbone. Next, quantization is employed during training. Finally, a special structure named bottleneck is applied in the model’s classification and regression part, reducing the number of parameters and the computational effort of the model.

4.8. Discussion

The results shown in the above datasets illustrate that our model gains excellent performance in certain scenarios, such as deformation challenges, scale variation challenges, background clutter challenges, etc., which accurately suggests that our module is effective. Our PAM ensures consistency between the tracking phase and training phase, leading to more accurate target finding in some turbulent scenarios.
In addition, the scores of the RT test on the VOT dataset indicate that our strategy of adopting different networks in the training and tracking phases is efficient, and the inference speed of classification and regression is significantly improved. Moreover, an unexpectedly exciting result arose: our results are much better on the UAV dataset than other anchor-free trackers.
However, the performance of our tracker in fast motion is not satisfactory compared to SiamCAR since we use a few frames close to each other as the target score for fitting. Similarly, it also impacts the robustness of tracking, which interprets that our R-score is poor compared to SiamCAR in the VOT baseline test. We will directly increase the range sizes between the frames to minimize errors in future experiments.
In conclusion, our experimental results can support that our PAM module and Acc branch are very effective in Single Object Tracking, and the PAM could significantly increase the performance of the anchor-free tracking network.

5. Conclusions and Future Work

A novel tracker named PACR is proposed to solve the inconsistency of object location rules between the tracking phase and training phase for Single Object Tracking. Our basic tracking framework is inspired by the Siamese-based networks. The PAM is proposed in our new tracker, and the output of the PAM will be sent to the Acc branch to adjust and fine-tune the result of the classification branch and regression branch, especially in some scenarios such as occlusion and deformation. Our tracker can achieve good results on OTB100, VOT2020, and LaSOT. Moreover, compared with other UAV trackers such as TCT and AutoTrack, our tracker achieves the best performance on DTB70 and UAV123.
In the future, we plan to apply a novel encoder–decoder architecture to our Acc branch, with the ConvNeXt being the backbone to extract the feature. During the ablation study, we tried to use only the PAM or the Acc branch in our tracker, and the results were worse than the comic SiamCAR. This is because the ideal result of the Acc branch requires the acceleration of the target.

Author Contributions

Conceptualization, D.L. and Q.W.; methodology, D.L. and H.C.; software, H.C. and Y.Z.; validation, H.C., Y.Z. and Y.X.; formal analysis, D.L. and H.C.; investigation, D.L.; resources, Y.X.; data curation, Q.W.; writing—original draft preparation, D.L. and H.C.; writing—review and editing, D.L. and Q.W.; visualization, H.C.; supervision, D.L.; project administration, Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC) grant number 61801340.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The code could be available at https://github.com/wernety/PACR, accessed on 17 February 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Chen, D.; Yang, T.; Hu, B.; Zhang, J. Action recognition based on object tracking and dense trajectories. In Proceedings of the IEEE International Conference on Automatica, Curico, Chile, 19–21 October 2016; pp. 1–5. [Google Scholar]
  2. Kiani Galoogahi, H.; Fagg, A.; Lucey, S. Learning background- aware correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1135–1143. [Google Scholar]
  3. Zhang, J.M.; Yuan, T.Y.; He, Y.Q.; Wang, J. A backgroundaware correlation filter with adaptive saliency-aware regularization for visual tracking. Neural Comput. Appl. 2022, 34, 6359–6376. [Google Scholar] [CrossRef]
  4. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-by-Detection with Kernels. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
  5. João, F.; Henriques, R.C.; Martins, P.; Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar]
  6. Danelljan, M.; Häger, G.; Khan, F.S.; Felsberg, M. Learning Spatially Regularized Correlation Filters for Visual Tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  7. Li, F.; Tian, C.; Zuo, W.; Zhang, L.; Yang, M.-H. Learning Spatial-Temporal Regularized Correlation Filters for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4904–4913. [Google Scholar]
  8. Huang, Z.; Fu, C.; Li, Y.; Lin, F.; Lu, P. Learning Aberrance Repressed Correlation Filters for Real-Time UAV Tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2891–2900. [Google Scholar]
  9. Li, Y.; Fu, C.; Ding, F.; Huang, Z.; Lu, G. AutoTrack: Towards High-Performance Visual Tracking for UAV with Automatic Spatio-Temporal Regularization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11920–11929. [Google Scholar]
  10. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H.S. Fully-Convolutional Siamese Networks for Object Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 850–965. [Google Scholar]
  11. He, A.; Luo, C.; Tian, X.; Zeng, W. A Twofold Siamese Network for Real-Time Object Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4834–4843. [Google Scholar]
  12. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High Performance Visual Tracking with Siamese Region Proposal Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 8971–8980. [Google Scholar]
  13. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Distractor-aware Siamese Networks for Visual Object Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Cham, Switzerland, 5–9 October 2018; pp. 103–119. [Google Scholar]
  14. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4277–4286. [Google Scholar]
  15. Xu, Y.; Wang, Z.; Li, Z.; Yuan, Y.; Yu, G. SiamFC++: Towards Robust and Accurate Visual Tracking with Target Estimation Guidelines. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12549–12556. [Google Scholar]
  16. Fu, Z.; Liu, Q.; Fu, Z.; Wang, Y. STMTrack: Template-free Visual Tracking with Space-time Memory Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13769–13778. [Google Scholar]
  17. Cui, Y.; Jiang, C.; Wang, L.; Wu, G. MixFormer: End-to-End Tracking with Iterative Mixed Attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13598–13608. [Google Scholar]
  18. Cao, Z.; Huang, Z.; Pan, L.; Zhang, S.; Liu, Z.; Fu, C. TCTrack: Temporal Contexts for Aerial Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  19. Xie, F.; Wang, C.; Wang, G.; Cao, Y.; Yang, W.; Zeng, W. Correlation-Aware Deep Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8741–8750. [Google Scholar]
  20. Gao, S.; Zhou, C.; Ma, C.; Wang, X.; Yuan, J. AiATrack: Attention in Attention for Transformer Visual Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Gothenburg, Sweden, 21-26 August 2022; pp. 146–164. [Google Scholar]
  21. Ye, B.; Chang, H.; Ma, B. Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework. In Proceedings of the European conference on computer vision (ECCV), Gothenburg, Sweden, 21–26 August 2022; Volume 13682, pp. 341–357. [Google Scholar]
  22. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual Object Tracking Using Adaptive Correlation Filters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
  23. Fu, C.; Li, B.; Ding, F.; Lin, F.; Lu, G. Correlation Filters for Unmanned Aerial Vehicle-Based Aerial Tracking: A Review and Experimental Evaluation. IEEE Geosci. Remote Sens. Mag. March 2022, 10, 125–160. [Google Scholar] [CrossRef]
  24. Shen, Z.; Dai, Y.; Rao, Z. CFNet: Cascade and Fused Cost Volume for Robust Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13901–13910. [Google Scholar]
  25. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. Int. J. Comput. Vis. 2020, 128, 642–656. [Google Scholar] [CrossRef] [Green Version]
  26. Zhou, X.; Zhuo, J.; Krähenbühl, P. Bottom-up Object Detection by Grouping Extreme and Center Points. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 850–859. [Google Scholar]
  27. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  28. Zhang, Z.; Peng, H.; Fu, J.; Li, B.; Hu, W. Ocean: Object-aware Anchor-free Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Germany, 23–28 August 2020; pp. 771–787. [Google Scholar]
  29. Cui, Y.; Guo, D.; Shao, Y.; Wang, Z.; Shen, C.; Zhang, L.; Chen, S. Joint Classification and Regression for Visual Tracking with Fully Convolutional Siamese Networks. Int. J. Comput. Vis. 2022, 130, 550–566. [Google Scholar] [CrossRef]
  30. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the ICLR, Vienna, Austria, 4 May 2021. [Google Scholar]
  31. Lin, L.; Fan, H.; Xu, Y.; Ling, H. Swintrack: A simple and strong baseline for transformer tracking. arXiv 2022, arXiv:2112.00995.2021. [Google Scholar]
  32. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 548–558. [Google Scholar]
  33. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. PVT v2: Improved Baselines with Pyramid Vision Transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  34. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar]
  35. Danelljan, M.; Bhat, G.; Paul, M.; Paudel, D.P.; Yu, F.; Van Gool, L. Transforming Model Prediction for Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8721–8730. [Google Scholar]
  36. Chen, X.; Bin, Y.; Zhu, J.; Wang, D.; Yang, X.; Lu, H. Transformer Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8122–8131. [Google Scholar]
  37. Zhang, J.; Xie, X.; Zheng, Z.; Kuang, L.-D.; Zhang, Y. SiamOA: Siamese offset-aware object tracking. Neural Comput. Appl. 2022, 34, 22223–22239. [Google Scholar] [CrossRef]
  38. Zhang, H.; Zu, K.; Lu, J.; Zou, Y.; Meng, D. EPSANet: An Efficient Pyramid Split Attention Block on Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 23556–23565. [Google Scholar]
  39. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the NIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  40. Chen, X.; Li, D.; Zou, Q. Exploiting Acceleration of the Target for Visual Object Tracking. IEEE Access 2021, 9, 73818–73825. [Google Scholar] [CrossRef]
  41. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  42. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vision 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  43. Fan, H.; Bai, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Harshit; Huang, M.; Liu, J.; et al. LaSOT: A High-quality Large-scale Single Object Tracking Benchmark. Int. J. Comput. Vis. 2020, 129, 439–461. [Google Scholar] [CrossRef]
  44. Real, E.; Shlens, J.; Mazzocchi, S.; Pan, X.; Vanhoucke, V. Youtube-boundingboxes: A large high-precision human-annotated data set for object detection in video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5296–5305. [Google Scholar]
  45. Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for uav tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 445–461. [Google Scholar]
  46. Li, S.; Yeung, D.-Y. Visual Object Tracking for Unmanned Aerial Vehicles: A Benchmark and New Motion Models. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 1–7. [Google Scholar]
  47. Matej, K.; Ales, L.; Jiri, M.; Michael, F.; Roman, P. The eighth visual object tracking vot2020 challenge results. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 547–601. [Google Scholar]
  48. Wu, Y.; Lim, J.; Yang, M.H. Object Tracking Benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Ross, D.; Lim, J.; Lin, R.S.; Lim, J. Incremental learning for robust visual tracking. Int. J. Comput. Vis. 2008, 77, 125r141. [Google Scholar] [CrossRef]
  50. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. Atom: Accurate tracking by overlap maximization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4660–4669. [Google Scholar]
  51. Tang, Z.; Luo, L.; Xie, B.; Zhu, Y.; Zhao, R.; Bi, L.; Lu, C. Automatic sparse connectivity learning for neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–15. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The current method of determining the tracking target is shown by the orange dash line (all points close to the tracking center are positive samples), which contains a lot of background information that will cause the tracker to perform wrong learning during the training phase. Pixel-attention encodes each pixel and calculates the weight, suppressing the small correlation score when generating positive sample labels.
Figure 1. The current method of determining the tracking target is shown by the orange dash line (all points close to the tracking center are positive samples), which contains a lot of background information that will cause the tracker to perform wrong learning during the training phase. Pixel-attention encodes each pixel and calculates the weight, suppressing the small correlation score when generating positive sample labels.
Mathematics 11 01406 g001
Figure 2. The entire network of PACR. The dimension of every model is shown near the model. We use concat and transposed conv to combine the feature extracted from the last three layers of the backbone. The PAM is the pixel-attention model.
Figure 2. The entire network of PACR. The dimension of every model is shown near the model. We use concat and transposed conv to combine the feature extracted from the last three layers of the backbone. The PAM is the pixel-attention model.
Mathematics 11 01406 g002
Figure 3. Detailed illustration of the tracking phase. The first rectangle is the picture of the search region and the template picture. The top four pictures indicate the acc score, cls score, cen score, and Hanning window, respectively. The bottom four pictures surrounded by the blue dot-line indicate the reg score.
Figure 3. Detailed illustration of the tracking phase. The first rectangle is the picture of the search region and the template picture. The top four pictures indicate the acc score, cls score, cen score, and Hanning window, respectively. The bottom four pictures surrounded by the blue dot-line indicate the reg score.
Mathematics 11 01406 g003
Figure 4. The framework of the pixel-attention module. The input is the correlation of the template feature and the search feature.
Figure 4. The framework of the pixel-attention module. The input is the correlation of the template feature and the search feature.
Mathematics 11 01406 g004
Figure 5. The upgraded pixel-attention module, compared with the old version. The input of this module changed to use both features of the previous frame and features of the current frame.
Figure 5. The upgraded pixel-attention module, compared with the old version. The input of this module changed to use both features of the previous frame and features of the current frame.
Mathematics 11 01406 g005
Figure 6. The result of tracking the occluded target compared with other trackers. The yellow line is our tracker, the green line is SiamCAR, and the light-blue line is the STMTracker.
Figure 6. The result of tracking the occluded target compared with other trackers. The yellow line is our tracker, the green line is SiamCAR, and the light-blue line is the STMTracker.
Mathematics 11 01406 g006
Figure 7. The result on different dimensions of OTB100 using OPE to evaluate.
Figure 7. The result on different dimensions of OTB100 using OPE to evaluate.
Mathematics 11 01406 g007
Figure 8. The evaluation results on the UAV123 dataset compared with some anchor-free trackers such as STMT, SiamCAR, Ocean, etc. (a) The precision points indicate the distance from the center point of the ground truth to the center point of the calculation by the trackers; (b) the success points indicate the overlap between the ground truth and the prediction.
Figure 8. The evaluation results on the UAV123 dataset compared with some anchor-free trackers such as STMT, SiamCAR, Ocean, etc. (a) The precision points indicate the distance from the center point of the ground truth to the center point of the calculation by the trackers; (b) the success points indicate the overlap between the ground truth and the prediction.
Mathematics 11 01406 g008
Figure 9. The result plot of DTB70 compared with TCTrack, SiamCAR, ATOM, etc. (a) The success points indicate the overlap between the ground truth with the prediction; (b) the precision points indicate the distance from the ground truth to the prediction.
Figure 9. The result plot of DTB70 compared with TCTrack, SiamCAR, ATOM, etc. (a) The success points indicate the overlap between the ground truth with the prediction; (b) the precision points indicate the distance from the ground truth to the prediction.
Mathematics 11 01406 g009
Table 1. The result of the VOT2020 dataset of PACR compared with other trackers. RT means real-time challenges. IVT (Incremental Visual Tracking), ATOM (Accurate Tracking by Overlap Maximization).
Table 1. The result of the VOT2020 dataset of PACR compared with other trackers. RT means real-time challenges. IVT (Incremental Visual Tracking), ATOM (Accurate Tracking by Overlap Maximization).
AREAOAUCA(RT)R(RT)EAO(RT)
IVT [49]0.3450.2440.0920.0960.3490.2290.089
KCF0.4070.4320.1540.1780.1540.4060.434
SiamRPN++0.4430.6720.2440.2290.4430.6730.244
ATOM figure [50]0.4620.7340.2710.3780.4400.6870.237
SiamFC0.4180.5020.1790.2290.4220.4790.172
SiamCAR0.4490.7320.2730.3570.4490.7320.272
PACR0.4560.7230.2620.3740.5060.6800.280
Table 2. The result of the evaluation on LaSOT.
Table 2. The result of the evaluation on LaSOT.
ECOSiamRPN++AToMTCTrackSiamCarPACR
Norm-precision0.3380.5690.5760.4840.6100.674
Auc0.3240.4960.5150.4350.5160.581
Precision0.3010.4910.5050.4140.5240.587
Table 3. The model complexity of each module which is calculated by PyTorch.
Table 3. The model complexity of each module which is calculated by PyTorch.
MemoryFLOPSMAddMemR + W
Backbone306.11 M44.66 G89.27 G772.59 M
Neck5.64 M883.19 M1.76 G25.06 M
PAM2.18 M414.9 M830.61 M50.5 M
ACRhead14.67 M5.92 G2.96 G40.07 M
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, D.; Chai, H.; Wei, Q.; Zhang, Y.; Xiao, Y. PACR: Pixel Attention in Classification and Regression for Visual Object Tracking. Mathematics 2023, 11, 1406. https://doi.org/10.3390/math11061406

AMA Style

Li D, Chai H, Wei Q, Zhang Y, Xiao Y. PACR: Pixel Attention in Classification and Regression for Visual Object Tracking. Mathematics. 2023; 11(6):1406. https://doi.org/10.3390/math11061406

Chicago/Turabian Style

Li, Da, Haoxiang Chai, Qin Wei, Yao Zhang, and Yunhan Xiao. 2023. "PACR: Pixel Attention in Classification and Regression for Visual Object Tracking" Mathematics 11, no. 6: 1406. https://doi.org/10.3390/math11061406

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop