Next Article in Journal
A Fortunate Refining Trip Recommendation Model
Previous Article in Journal
Improved Belief Propagation List Decoding for Polar Codes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dim and Small Target Tracking Using an Improved Particle Filter Based on Adaptive Feature Fusion

1
Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
China Astronaut Research and Training Center, Beijing 100094, China
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(15), 2457; https://doi.org/10.3390/electronics11152457
Submission received: 13 July 2022 / Revised: 4 August 2022 / Accepted: 5 August 2022 / Published: 7 August 2022
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Particle filters have been widely used in dim and small target tracking, which plays a significant role in navigation applications. However, their characteristics, such as difficulty of expressing features for dim and small targets and lack of particle diversity caused by resampling, lead to a considerable negative impact on tracking performance. In the present paper, we propose an improved resampling particle filter algorithm based on adaptive multi-feature fusion to address the drawbacks of particle filters for dim and small target tracking and improve the tracking performance. We first establish an observation model based on the adaptive fusion of the features of the weighted grayscale intensity, edge information, and wavelet transform. We then generate new particles based on residual resampling by combining the target position in the previous frame and the particles in the current frame with higher weights, with the tracking accuracy and particle diversity improving simultaneously. The experimental results demonstrate that our proposed method achieves a high tracking performance with a distance accuracy of 77.2% and a running speed of 106 fps, respectively, meaning that it will have a promising prospect in dim and small target tracking applications.

1. Introduction

Video processing technologies such as target tracking [1,2], target detection [3,4], and moving target segmentation [5,6] have made great progress, with related applications continuing to emerge. Among them, target tracking has been widely used in intelligent video surveillance, modern military, and intelligent visual navigation equipment. A set of conventional methods have been proposed to meet the application requirements, including particle filters (PF) [7], kernelized correlation filters (KCF) [8], efficient convolution operators (ECO) [9], and spatial-temporal regularized correlation filters (STRCF) [10]. Dim and small target tracking is an important branch of target tracking, which plays a key role in military, aviation, and aerospace fields such as image matching guidance and reconnaissance [2]. Dim and small target tracking possesses more stringent requirements for trackers because the targets occupy only a small number of pixels in the image (the target size is generally about 2 × 2), which means that there is a lack of feature information, such as shape and texture, and a sensitivity to noise [11].
Recently, scholars have proposed many tracking algorithms to meet the requirements of the dim and small target tracking, which can be divided into two main categories including correlation filters-based methods [12,13,14] and PF-based methods [15,16,17,18,19]. KCF has received the most attention in the methods based on correlation filters for dim and small target tracking, which is achieved by establishing a discriminator based on the correlation operator with a kernel function. Qian K. et al. proposed an anti-interference small target tracking algorithm [12], which combines KCF with a detection model (KCFD), making a robust tracking result on image sequences with complex backgrounds. Zhang L. et al. proposed an infrared small target tracking algorithm consisting of a discriminator based on KCF and a predictor based on the least-square trajectory prediction [13], which made full use of the continuity and direction of the target motion and can robustly track the target with shot-term occlusions. Kou Z. et al proposed a method based on target spatial distribution with improved KCF for infrared small target tracking [14], which considers the importance of intensity features to infrared targets and different regions to calculate the gray distribution weighting function of the target, solving the problems of target occlusion and drift. However, the above three methods based on KCF cannot achieve a higher tracking performance on the image sequences with a fast-moving target. A considerable number of previous studies on dim and small target tracking focus on particle filtering [15], which expresses the various motion states that the target possibly owns by the posteriori probability distribution estimated by a group of weighted particles, and it is a favorable method for solving nonlinear and non-Gaussian problems, fitting for dim and small target tracking. Conventional particle filtering has the following drawbacks during the dim and small target tracking task. First, the extracted features cannot explicitly represent the target and cause a tracking drift. Second, the conventional resampling method typically leads to a lack of particle diversity during the resampling process, which causes tracking performance degradation. Consequently, a considerable number of methods have been proposed to address the above drawbacks of particle filters for dim and small target tracking. Fan X. et al. proposed a particle filter method for adaptive template updating, combining the neighborhood motion model and the grayscale probability graph [16], which enhances the ability of dim and small target tracking. Ji E. et al. improved the mean-shift based particle filter tracking algorithm according to multi-feature fusion for small target tracking [17], which fuses a high-frequency histogram, fractal, and the energy of an infrared small target to improve the accuracy of the tracking process. Wang Y. et al. integrated the genetic resampling method into the particle filter method for small moving target tracking [18], which avoids particle degeneracy and guarantees particle diversity. Tian M. et al. proposed a track-before-detect method based on the spring model firefly algorithm optimization particle filter (SFA-PF-TBD) for dim small target tracking [19], making the distribution of particles more reasonable. However, these methods based on PF have a higher computational cost and may be prone to fall into local optimal values.
In the present paper, we propose a tracking algorithm based on adaptive feature fusion and an improved resampling particle filter, which is used to solve the two problems mentioned above of a particle filter for dim and small target tracking, namely a lack of the target features and the particle diversity, while reducing the computational cost of the algorithm. We first perform adaptive fusion for three features (grayscale intensity, edge information, and wavelet transform) based on the similarity of the features between the candidate region and the target. We then solve the problem of a lack of particle diversity by improving the residual resampling algorithm and generating some new particles that are meaningful for subsequent frames. In comparison to the methods PF [7], KCFD [12], and SFA-PF-TBD [19], our proposed method achieves a reasonable and robust tracking performance for six different dim and small target image sequences. The highlights and contributions of the present paper are as follows:
  • Our method adaptively fuses three kinds of features to express the dim and small targets more accurately, making it robustly track the targets in various complex scenes.
  • Our improved resampling method addresses the lack of particle diversity with a lower computational complexity, which makes a good balance between the tracking performance and computational cost.
This paper is organized as follows: Section 2 illustrates the contextual information for the particle filter algorithm; Section 3 exhibits the details of our method; Section 4 presents the experimental results and comparisons to other methods; and, finally, Section 5 concludes the paper and outlines the future directions.

2. Background Information

A particle filter is a nonlinear filtering algorithm based on the Monte Carlo method for Bayesian estimations, which is used to solve the complex multi-integral calculation problem in Bayesian filtering. Additionally, the central aim concerning a particle filter is to estimate the probability density with a set of weighted random samples (that is, particles) and then to perform a weighted summation on these particles to obtain the minimum variance estimate.
The state and observation equations of the target tracking system are set as follows:
X t = f ( X t 1 ) + Q t , Y t = h ( X t ) + R t
where X t and Y t represent the state and observation values of the system at time t, respectively; f ( ) and h ( ) represent the system state transition function and observation function, respectively; and Q t and R t represent the system process noise and observation noise, respectively.
Bayesian filtering is generally divided into two parts: the prediction and update steps. The prediction step estimates the probability density of the current moment according to the posterior probability density of the previous moment:
p ( X t | Y 1 : t 1 ) = p ( X t | X t 1 ) p ( X t 1 | Y 1 : t 1 ) d X t 1
The update step updates the posterior probability density of the current moment according to the estimated probability density and observed value of the current moment:
p ( X t | Y 1 : t ) = η [ p ( Y t | X t ) p ( X t | Y 1 : t 1 ) ]
where η = [ p ( Y t | X t ) p ( X t | Y 1 : t 1 ) d X t ] 1 . Then, calculate the expected value of the posterior probability p ( X t | Y 1 : t ) to obtain the final estimated state value:
X ^ t = X t p ( X t | Y 1 : t ) d X t
The particle filter samples the posterior probability density based on the Monte Carlo method. However, it is difficult to directly sample the posterior probability in the actual situation, so an importance function, q ( X t | Y 1 : t ) , is introduced, and Equation (4) is changed to the following:
X ^ t = X t p ( X t | Y 1 : t ) q ( X t | Y 1 : t ) q ( X t | Y 1 : t ) d X t
Then, we can apply the Monte Carlo method to sample q ( X t | Y 1 : t ) to obtain
X ^ t i = 1 N w i t X t i
where the particle weight w i k is defined as follows:
w i t = p ( X t i | Y 1 : t ) q ( X t i | Y 1 : t )
Transform p ( X t | Y 1 : t ) and q ( X t | Y 1 : t ) to
p ( X t | Y 1 : t ) p ( Y t | X t ) p ( X t | X t 1 ) p ( X t 1 | Y 1 : t 1 )
q ( X t | Y 1 : t ) = q ( X t | X t 1 , Y 1 : t ) q ( X t 1 | Y 1 : t 1 )
Based on Equations (7)–(9), we can obtain the recurrence equation of the weight w i k :
w i t w i t 1 p ( Y t | X i t ) p ( X i t | X i t 1 ) q ( X i t | X i t 1 , Y 1 : t )
Equations (6), (9) and (10) constitute the main framework of particle filtering. Firstly, the importance function of the current moment is calculated according to the importance function at time t − 1 and Equation (9). Then, using Equation (10) allows us to obtain the particle weight w i k . Finally, according to Equation (6), the estimated value of the state in the current moment is obtained.

3. Principle of the Proposed Method

3.1. Feature Extraction

Although the gray histogram feature can clearly express the grayscale distribution of the image, it lacks the description of the spatial position distribution, so it can be considered to obtain the gray histogram feature by combining the weighted position information [20]. We introduced a kernel function k ( r ) :
k ( r ) = { 1 r 2 , r < 1 0 , r 1
which assigns different k values based on the distance r. The longer the distance r, the small the k value. Then, the gray feature with the position distribution information f g r a y can be expressed as follows:
f g r a y = α i = 1 n k ( | | x 0 x i | | s ) δ [ b ( x i ) m ]
where n represents the number of pixels in the target area; m is the gray value range [ 0 , 255 ] ; x 0 is the position coordinate of the center pixel; x i is the position coordinate of the i-th pixel; b ( x i ) is the gray level of the pixel x i in the obtained histogram grade value; s = H x 2 + H y 2 ( H x and H y are the half-width and half-height of the area rectangle); and α is the normalization coefficient defined as follows.
α = [ i = 1 n k ( | | x 0 x i | | s ) ] 1
The image edge information also plays a significant role in the feature of the image, which is a collection of pixels where the distribution of the image features, such as pixel gray value and texture, are discontinuous or where there is a step change. Image edge features have a certain robustness towards illumination changes and rotations, which can make up for the defects of gray features. A high number of methods have been developed to calculate the edge information [21], which can express the target in which we are also interested. In this paper, we chose the Sobel operator to establish the edge histogram based on the edge gradient amplitude.
The Sobel operator performs pixel-by-pixel convolution of the image to be processed and the template in the form of a sliding window to obtain edge information. We choose the Sobel operator template in four directions (horizontal S x , vertical S y , 45 degrees S 45 ° , and 135 degrees S y ) to extract image edge information in all aspects. The four operator templates are as follows:
S x = [ 1 ,     2 ,     1 0 ,     0 ,     0 1 , 2 , 1 ] , S y = [ 1 ,     0 ,   1 2 ,     0 ,   2 1 ,     0 ,   1 ] , S 45 ° = [ 2 ,     1 ,     0 1 ,     0 , 1 0 , 1 , 2 ] , S 135 ° = [ 0 ,     1 ,     2 1 ,     0 ,     1 2 ,     1 ,     0 ]
Convolve the four templates S x , S y , S 45 ° , S 135 ° with the original image I pixel by pixel to obtain the edge gradient values in four directions:
G x = S x I , G y = S y I , G 45 ° = S 45 ° I , G 135 ° = S 135 ° I
where represents pixel-wise convolution. We then calculate the total edge gradient magnitude G by synthesizing the edge gradient values in the four directions:
G = G x 2 + G y 2 + G 45 ° 2 + G 135 ° 2
Finally, the edge histogram f e d g e can be extracted based on the kernel function presented in Equations (11) and (12).
To better express the target features, we considered combining the frequency domain features of the image. We chose the wavelet features because the wavelet transform has good time–frequency localization characteristics, which is convenient for adjusting the filter direction and fundamental frequency bandwidth, so as to better take into account the resolution of the spatial and frequency domains [22]. Moreover, the wavelet feature is insensitive to illumination changes and can tolerate target rotation and deformation to a certain extent.
The two-dimensional wavelet function H ( x , y ) can be expressed as follows:
H ( x , y ) = 1 2 π σ x σ y exp ( 1 2 ( x 2 σ x 2 + y 2 σ y 2 ) ) cos ( ω x )
where σ x and σ y are the standard deviations on the x and y axes, respectively, which determine the size of the filtering area. ω is the baseband bandwidth. By convolving the original image I with H ( x , y ) , we can obtain the wavelet transform W of I:
W ( x , y ) = H ( x , y ) I ( x , y )
Here, represents the convolution operation. Similarly, Equations (11) and (12) can be used to extract the histogram f w a v e l e t of W.

3.2. Feature Fusion and Model Establishment

We set the fusion weight by comparing the correlation c i between the three features of the initial target f _ t i and the three features of the current target f _ c i to perform adaptive feature fusion, where i = ( g r a y , e d g e , w a v e l e t ) . The correlation calculation formula is as follows:
c i ( f _ t i , f _ c i ) = j J [ f _ t i ( j ) f _ t i ¯ ] [ f _ c i ( j ) f _ c i ¯ ] j J [ f _ t i ( j ) f _ t i ¯ ] 2 j J [ f _ c i ( j ) f _ c i ¯ ] 2
where J represents the total number of feature histograms, f _ t i ¯ and f _ c i ¯ are the mean of f _ t i and f _ c i , respectively. We then obtained the correlation c i of two sets of gray features, two sets of edge features, and two sets of wavelet features ( c g r a y , c e d g e , and c w a v e l e t , respectively) based on Equation (19), and converted the correlation to the weight according to Equation (20):
v i = 0.5 ( c i + 1 ) i 0.5 ( c i + 1 ) , i = ( g r a y , e d g e , w a v e l e t )
Then, the adaptive fusion feature is expressed as follows:
f = i v i f i , i = ( g r a y , e d g e , w a v e l e t )
In the process of particle filter target tracking, it is necessary to update the estimate by using the observation value in the current moment. In the present paper, we chose to update the similarity between the fusion histogram feature f p of the target model and the fusion histogram feature f q of the candidate region. The similarity measure uses the Bhattacharyya distance d :
d = 1 ρ ( f p , f q )
where ρ ( f p , f q ) = i M f p ( i ) f q ( i ) and M is the histogram dimension. The larger ρ is, the more similar the candidate region histogram is to the target model histogram, that is, the candidate region is likely to be the estimated target position, and the particle will be given a higher weight.

3.3. Improved Resampling Particle Filter Algorithm

The resampling method generally adopted in the conventional particle filter algorithm is to simply copy the particles with high weights and to discard the particles with low weights, which will cause some particles with high weights to be sampled multiple times, resulting in a lack of samples and missing particle diversities [1]. In the current paper, based on residual resampling, we generated new particles according to high-weight particles in the current moment and the target position of the previous frame to solve the problem of a lack of particle samples, and the newly added particles can also track subsequent frames better.
First, we defined an effective particle number N e f f to judge whether the particles were degenerated, which is calculated based on weights of all particles and can clearly reflect the weight distribution of the entire particle set. N e f f is defined as follows:
N e f f = [ i = 1 N ( w k i ) 2 ] 1
where N is the total number of particles. The smaller the number of effective particles N e f f , the more uneven distribution of particle weights, that is, there may be a large number of particles with small weights and a few particles with large weights, which means that particle diversity is missing.
We set a threshold N t h ( N t h = 2 3 N in this paper); if N e f f N t h , we assumed that the particles had a certain diversity and no resampling was required; if N e f f < N t h , it was considered that the particles degenerated, and residual resampling needed to be performed. Residual resampling consists of two parts as follows:
(1). We copied the valid particles, and the number of copies was determined by the weight of the particles. For the particle set [ P k _ o l d ( i ) , w k _ o l d ( i ) ] , we preserved n k ( i ) = N w k _ o l d ( i ) particles and then obtained set [ P k _ r ( i ) , w k _ r ( i ) ] , where represents the rounding operation, the corrected weights w k _ r ( i ) = ( N w k _ o l d ( i ) n k ( i ) ) / N , and N = i = 1 N n k ( i ) .
(2). When the number of residual particles M = N N was greater than 0, we randomly resampled the particles [ P k _ r ( i ) , w k _ r ( i ) ] and sorted them according to their weight to obtain the particle set [ P k ( i ) , w k ( i ) ] .
The positions close to the particles with higher weights are likely to be the potential position of the target in the subsequent frame, which may be the particles useful for solving the problem of a lack of particle diversity and for facilitating the tracking of the subsequent frame. Therefore, we chose the first J particles P k ( j ) = ( x k ( j ) , y k ( j ) ) , ( j = 1 , , J ) in [ P k ( i ) , w k ( i ) ] with high weights and adopted the points with a distance Δ d ( Δ d = 2 , in this paper) around P k ( j ) = ( x k ( j ) , y k ( j ) ) , ( j = 1 , , J ) as the new particle candidate positions, as follows:
P k _ n e w ( j ) = ( [ x k ( j ) ; 1 ] Δ x , [ y k ( j ) ; 1 ] Δ y )
where Δ x and Δ y represent the coordinates of the points around a particle with a distance of Δ d in eight directions, and Δ x , Δ y are defined as Equation (25). Therefore, according to Equation (24), the points around the particles with the larger weights in the particle set P k ( j ) = ( x k ( j ) , y k ( j ) ) , ( j = 1 , , J ) can be extracted as the new candidate particles.
Δ x = [ 1 1 1 1 1 1 1 1 Δ d Δ d 0 0 Δ d Δ d Δ d Δ d ] Δ y = [ 1 1 1 1 1 1 1 1 0 0 Δ d Δ d Δ d Δ d Δ d Δ d ]
We can obtain a total of l candidate particles by eliminating the repeated values appearing in P k _ n e w ( j ) and can calculate the distance d o between the positions of these l candidate particles and the estimated target position at the last moment. The first 0.2 N candidate particles with the closest distance were used to substitute the 0.2 N particles with low weights in P k ( i ) , and a new particle set was obtained. This not only can generate new particles to solve the problem of a lack of diversity of the traditional resampling particles but also can gather more effective particles at the potential positions of the target in the subsequent frame to facilitate subsequent tracking.

3.4. Algorithm Process

The pseudo code of the algorithm in this paper is shown in Algorithm 1.
Algorithm 1. The pseudo code of the algorithm we proposed.
Input: image sequences I and target position in initial frame.
Output: the target position Pos in subsequent frames.
% Initial frame. Initialization
Extract f g r a y ,   f e d g e ,   f w a v e l e t ; and fuse them to get f p ;
Initialize N particles and perform importance sampling.
% Subsequent frames. Tracking
for frame=2: length(I):
   Extract f g r a y , f e d g e , f w a v e l e t at current frame;
   Fuse f g r a y , f e d g e , f w a v e l e t to get f q according to Equations (19)–(21);
    d ← Equation (22);
    w i k ← Equation (10);
   Pos← Equation (6);
   % Resampling
   if N e f f N t h :
      continue;
   else:
      Perform residual resampling;
      Generate new particles according to Equation (24).
   end if
end for

4. Experimental Results and Analysis

To evaluate the tracking performance of the proposed method relative to others, the results of the tracking experiments were compared to the methods of PF [7], KCFD [12], and SFA-PF-TBD [19]. The experimental data adopted the data set published by Hui B. et al. [23], in which six groups of sequences, data5, data8, data16, data18, data19, and data20, were chosen for the tracking experiment. The resolutions of the images of six sequences were 256 × 256, and more specific descriptions about the six sequences are shown in Table 1. For a fair comparison of the tracking performance and computational cost, we performed experiments using the four methods by running them on a laptop with 2.40 GHz Intel i5–1135G7 and 16GB RAM, and with MATLAB 2018a.

4.1. Qualitative Analysis

Figure 1 shows the dynamic process of target tracking in the super long-time sequence (data5), which is used to verify the ability of the trackers to track dim targets affected by noise for a long period of time. As shown in Figure 1a, KCFD and PF deviate from the target due to the weak target, in frame 127. Additionally, we can observe in Figure 1b that methods PF, SFA-PF-TBD, and KCFD have completely lost the target before reaching frame 700, while our proposed method can better estimate the target state and stably track the target for a longer period of time, as shown in Figure 1c.
Figure 2 shows the dynamic process of target tracking in the complex background sequence (data8). The images in this sequence have a complex background accompanied by objects with a grayscale similar to the target. As shown in Figure 2a, the target is obvious enough against a simple background so that all four methods can stably track the target before reaching frame 54. As shown in Figure 2b, in frame 203, objects similar to the target appear in the background, causing the KCFD method to fail in tracking the target. Similarly, in frame 366, the complicated background causes SFA-PF-TBD tracking to deviate from the target, as shown in Figure 2c. Our proposed method and PF can adapt to the complex backgrounds and robustly track the target.
Figure 3 shows the dynamic process of target tracking in the fast-moving sequence (data16). The target locations change rapidly at certain frames in data16, which requires the trackers to have the ability to rapidly retrieve the target after the target is lost. As shown in Figure 3a, the target moves suddenly and rapidly in frame 134, which causes all the methods to fail to track the target. Our proposed method and SFA-PF-TBD can rapidly retrieve the target at frame 140 and keep tracking the target successfully; PF and KCFD cannot identify and track the target, as shown in Figure 3b,c.
Figure 4 shows the dynamic process of target tracking in the fast-moving and complex background sequence (data18). As shown in Figure 4a, rapid changes in the target’s position due to camera micro-vibrations and the complex background in frame 22 cause the four methods to deviate from the target. However, the proposed method and PF immediately retrieved the target at frame 27, as shown in Figure 4b, and it can be observed in Figure 4c that a deviation is produced by PF in frame 169 due to the high similarity between the target and background, while the proposed method can effectively track fast-moving targets with complex backgrounds.
Figure 5 shows the dynamic process of target tracking in the long-time and complex background sequence (data19). As shown in Figure 5a,b, PF and KCFD deviate from the target and fail in their tracking processes because of the complicated background in frame 538 and frame 751. However, our method and SFA-PF-TBD can effectively track the target for a long period of time. At frame 800, SFA-PF-TBD tracking fails because the target becomes weak, while our method only has a slight deviation and can still continue to track the target, as shown in Figure 5c.
Figure 6 shows the dynamic process of target tracking in the background change and target rotation sequence (data20). SFA-PF-TBD tracking failed due to the background change (from sky to mountain) in frame 238, as shown in Figure 6a. In frame 279, we can observe in Figure 6b,c that the target becomes faint due to the rotation, causing our method, SFA-PF-TBD, and PF to deviate from the target. However, when the target is recovered at frame 333, our method can continue to track the target, while others lose the target completely.

4.2. Quantitative Analysis

To clearly compare the tracking performance of the proposed algorithm and the other three algorithms, we composed the average distance accuracy plots and average success rate plots for six groups of sequences, as shown in Figure 7a,b, respectively. It can be observed from the figure that the algorithm presented in the present paper is superior to the other three algorithms in terms of the average distance accuracy and average success rate.
Table 2 lists the average center pixel error, average 20-pixel distance accuracy, average overlap rate, and average tracking speed for six groups of sequences for the four algorithms. By comparison, it can be observed that, in terms of the average center pixel error, the proposed algorithm has the lowest error, followed by the SFA-PF-TBD algorithm. For the average 20-pixel distance accuracy, the algorithm presented in the current paper can reach up to 77.2%, and the SFA-PF-TBD is 68.6%. In relation to the average overlap rate, because the target size is too small, the overall overlap rate is low. Among the four algorithms, the average overlap rate of the proposed algorithm is the best, which is 27%. We compared the computational cost by calculating the average running speed (time to process each image) of the algorithms on the same laptop. The PF algorithm has the highest tracking speed, which can reach 121 fps, meaning it has the minimum computational cost. Compared with SFA-PF-TBD and KCFD, the algorithm presented in the current paper has a lower computational cost that achieves a running speed of 106 fps. Therefore, on the whole, the algorithm proposed in this paper has a greater tracking performance.

4.3. Merits and Limitations

The proposed method achieves stable performances for dim and small target tracking on the data set with a sky background. The fusion feature method we proposed in the present study can express dim and small targets well, which make the distance accuracy and overlap rate of the proposed method superior to the others. Moreover, the proposed method achieves a good balance between the tracking performances and computational cost, when compared with the other methods.
However, the proposed method also has some limitations. First, for some of the scenes with more complex backgrounds, a louder noise, and considerable similarities between the objects and backgrounds, our method may not accurately estimate the target state, resulting in the tracking method failing. Second, the proposed method cannot process the long-term occlusion and disappearance problems well.

5. Conclusions

In the current paper, we propose an improved resampling particle filter algorithm based on feature fusion, which solves the two problems in particle filters for dim and small target tracking, namely feature expression of dim and small targets and lack of the particle diversity. The fused feature of grayscale, edge, and wavelet clearly express the dim small target in various complex scenes. The improved resampling particle filter solves the problem of lacking particle diversity in a particle filter based on residual resampling with some new meaningful particles introduced, which boosts the performance of a particle filter tracker while reducing the computational complexity. The experimental results demonstrate that our method achieves a strong tracking performance with a distance accuracy of 77.2% and an overlap rate of 27%, which can effectively and robustly track dim and small targets. Moreover, our proposed method achieves a frame rate of 106 fps with a good balance between the tracking performance and computational cost.
Although the proposed method can achieve a reasonable performance for six image sequences with sky backgrounds, it cannot track the target well in some of the more complex scenes, such as the images presenting a louder noise and long-term occlusions. The focus of future research should be to choose more reasonable features concerning adaptive fusion to improve the tracking performance for more complex scenes.

Author Contributions

Conceptualization, H.Z. (Haifeng Zhang) and H.W.; funding acquisition, H.W.; investigation, H.Z. (Hongbo Zhang); methodology, Y.H. and H.Z. (Hongbo Zhang); software, Y.H. and Y.C.; writing—original draft, Y.H.; writing—review and editing, Y.C. and H.Z. (Haifeng Zhang). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the West Light Foundation of the Chinese Academy of Sciences, grant No. XAB2021YN15.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, T.; Bolic, M.; Djuric, P.M. Resampling Methods for Particle Filtering: Classification, Implementation, and Strategies. IEEE Signal Process. Mag. 2002, 32, 70–86. [Google Scholar] [CrossRef]
  2. Bai, Y.; Yang, D.Z.; Yu, S.; Li, J.Y.; Huang, C. Overview on Infrared Dim Target Tracking. In Proceedings of the Seventh Asia Pacific Conference on Optics Manufacture and 2021 International Forum of Young Scientists on Advanced Optical Manufacturing (APCOM and YSAOM 2021), Shanghai, China, 28–31 August 2021; SPIE: Bellingham, WA, USA, 2022; Volume 12166, pp. 1808–1815. [Google Scholar]
  3. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  4. Dong, Y.; Du, B.; Zhang, L. Target Detection Based on Random Forest Metric Learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 1830–1838. [Google Scholar] [CrossRef]
  5. Ivanov, Y.; Peleshko, D.; Makoveychuk, O.; Izonin, I.; Malets, I.; Lotoshunska, N.; Batyuk, D. Adaptive Moving Object Segmentation Algorithms in Cluttered Environments. In Proceedings of the 2015 13th International Conference the Experience of Designing and Application of CAD Systems in Microelectronics (CADSM 2015), Lviv, Ukraine, 24–27 February 2015; pp. 97–99. [Google Scholar]
  6. Peleshko, D.; Ivanov, Y.; Sharov, B.; Izonin, I.; Borzov, Y. Design and Implementation of Visitors Queue Density Analysis and Registration Method for Retail Videosurveillance Purposes. In Proceedings of the 2016 IEEE First International Conference on Data Stream Mining & Processing (DSMP), Lviv, Ukraine, 23–27 August 2016; pp. 159–162. [Google Scholar]
  7. Das, S.; Kale, A.; Vaswani, N. Particle Filter with a Mode Tracker for Visual Tracking Across Illumination Changes. IEEE Trans. Image Process. 2011, 21, 2340–2346. [Google Scholar] [CrossRef] [PubMed]
  8. João, F.H.; Rui, C.; Pedro, M.; Jorge, B. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar]
  9. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ECO: Efficient Convolution Operators for Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6931–6939. [Google Scholar]
  10. Li, F.; Tian, C.; Zuo, W.; Zhang, L.; Yang, M.-H. Learning Spatial-Temporal Regularized Correlation Filters for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4904–4913. [Google Scholar]
  11. Eysa, R.; Hamdulla, A. Issues on Infrared Dim Small Target Detection and Tracking. In Proceedings of the 2019 International Conference on Smart Grid and Electrical Automation (ICSGEA), Xiangtan, China, 10–11 August 2019; pp. 452–456. [Google Scholar]
  12. Qian, K.; Rong, S.H.; Cheng, K.H. Anti-Interference Small Target Tracking from Infrared Dual Waveband Imagery. Infrared Phys. Technol. 2021, 118, 103882. [Google Scholar] [CrossRef]
  13. Zhang, L.; Liu, K.; Gao, L. Infrared Small Target Tracking in Complex Background Based on Trajectory Prediction. In Proceedings of the Eleventh International Conference on Graphics and Image Processing (ICGIP 2019), Hangzhou, China, 12–14 October 2019; SPIE: Bellingham, WA, USA, 2020; Volume 11373, pp. 476–484. [Google Scholar]
  14. Kou, Z.; Hamdulla, A. Infrared Small Target Tracking Based on Target Spatial Distribution with Improved Kernelized Correlation Filtering. Res. Sq. 2021. [Google Scholar] [CrossRef]
  15. Cai, J.; Huang, Y.; Li, P. Small Target Tracking in Background Using Saliency-Based Particle Filter. In Proceedings of the Proceedings 2018 Chinese Automation Congress, CAC 2018, Xi’an, China, 30 November–2 December 2018; pp. 1350–1354. [Google Scholar]
  16. Fan, X.; Xu, Z.; Zhang, J. Dim Small Target Tracking Based on Improved Particle Filter. Guangdian Gongcheng/Opto-Electronic Eng. 2018, 45, 170569-1–170569-10. [Google Scholar]
  17. Ji, E.; Gu, G.; Qian, W.; Bai, L.; Sui, X. Improved Particle Filtering Algorithm Based on the Multi-Feature Fusion for Small IR Target Tracking. In International Symposium on Photoelectronic Detection and Imaging 2011: Advances in Infrared Imaging and Applications; SPIE: Bellingham, WA, USA, 2011; pp. 437–445. [Google Scholar]
  18. Wang, Y.; Wang, X.; Shan, Y.; Cui, N. Quantized Genetic Resampling Particle Filtering for Vision-Based Ground Moving Target Tracking. Aerosp. Sci. Technol. 2020, 103, 105925. [Google Scholar] [CrossRef]
  19. Tian, M.; Chen, Z.; Wang, H.; Liu, L. An Intelligent Particle Filter for Infrared Dim Small Target Detection and Tracking. IEEE Trans. Aerosp. Electron. Syst. 2022, 9251, 1. [Google Scholar] [CrossRef]
  20. Comaniciu, D.; Ramesh, V.; Meer, P. Kernel-Based Object Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 564–577. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, C.; Li, Y.; Qiu, A. Comparison Research of Capability of Several Edge Detection Operators. In Proceedings of the International Conference on Industrial Technology and Management Science, Tianjin, China, 27–28 March 2015; Atlantis Press: Amsterdam, The Netherlands, 2015; pp. 795–798. [Google Scholar]
  22. He, C.; Zheng, Y.F.; Ahalt, S.C. Object Tracking Using the Gabor Wavelet Transform and the Golden Section Algorithm. IEEE Trans. Multimed. 2002, 4, 528–538. [Google Scholar]
  23. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Lin, J.; Su, H.; Jin, W.; Zhang, Y.; et al. A Dataset for Infrared Image Dim-Small Aircraft Target Detection and Tracking under Ground/Air Background. Sci. Data Bank 2019, 5, 12. [Google Scholar]
Figure 1. Tracking results of four methods for data5. (a) Results at frame 127; (b) results at frame 700; (c) results at frame 768.
Figure 1. Tracking results of four methods for data5. (a) Results at frame 127; (b) results at frame 700; (c) results at frame 768.
Electronics 11 02457 g001
Figure 2. Tracking results of four methods for data8. (a) Results at frame 54; (b) results at frame 203; (c) results at frame 366.
Figure 2. Tracking results of four methods for data8. (a) Results at frame 54; (b) results at frame 203; (c) results at frame 366.
Electronics 11 02457 g002
Figure 3. Tracking results of four methods for data16. (a) Results at frame 134; (b) results at frame 140; (c) results at frame 279.
Figure 3. Tracking results of four methods for data16. (a) Results at frame 134; (b) results at frame 140; (c) results at frame 279.
Electronics 11 02457 g003
Figure 4. Tracking results of four methods for data18. (a) Results at frame 22; (b) results at frame 27; (c) results at frame 169.
Figure 4. Tracking results of four methods for data18. (a) Results at frame 22; (b) results at frame 27; (c) results at frame 169.
Electronics 11 02457 g004
Figure 5. Tracking results of four methods for data19. (a) Results at frame 538; (b) results at frame 751; (c) results at frame 800.
Figure 5. Tracking results of four methods for data19. (a) Results at frame 538; (b) results at frame 751; (c) results at frame 800.
Electronics 11 02457 g005
Figure 6. Tracking results of four methods for data20. (a) Results at frame 238; (b) results at frame 279; (c) results at frame 333.
Figure 6. Tracking results of four methods for data20. (a) Results at frame 238; (b) results at frame 279; (c) results at frame 333.
Electronics 11 02457 g006
Figure 7. Average distance accuracy plot and success rate plot of four algorithms in six groups of sequences. (a) Precision plots and (b) success plots.
Figure 7. Average distance accuracy plot and success rate plot of four algorithms in six groups of sequences. (a) Precision plots and (b) success plots.
Electronics 11 02457 g007
Table 1. Detailed description of six groups of sequences.
Table 1. Detailed description of six groups of sequences.
SequenceTotal FramesTarget SizeSequence Characteristics
data530002 × 2Super long-time, weak target
data83992 × 2Weak target, complex background
data164995 × 5Move fast, from far to near
data185005 × 5Move fast, complex background
data1915992 × 2Long-time, complex background
data204002 × 2Weak target, target rotation
Table 2. Average center pixel error, 20-pixel distance accuracy, overlap rate, and running speeds of the four algorithms.
Table 2. Average center pixel error, 20-pixel distance accuracy, overlap rate, and running speeds of the four algorithms.
AlgorithmCenter Pixel Error/PixelsDistance Accuracy (20)Overlap RateSpeed/fps
Ours37.477.2%27.0%106
SFA-PF-TBD38.568.6%26.2%65
PF42.966.2%15.9%121
KCFD39.248.7%18.7%73
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huo, Y.; Chen, Y.; Zhang, H.; Zhang, H.; Wang, H. Dim and Small Target Tracking Using an Improved Particle Filter Based on Adaptive Feature Fusion. Electronics 2022, 11, 2457. https://doi.org/10.3390/electronics11152457

AMA Style

Huo Y, Chen Y, Zhang H, Zhang H, Wang H. Dim and Small Target Tracking Using an Improved Particle Filter Based on Adaptive Feature Fusion. Electronics. 2022; 11(15):2457. https://doi.org/10.3390/electronics11152457

Chicago/Turabian Style

Huo, Youhui, Yaohong Chen, Hongbo Zhang, Haifeng Zhang, and Hao Wang. 2022. "Dim and Small Target Tracking Using an Improved Particle Filter Based on Adaptive Feature Fusion" Electronics 11, no. 15: 2457. https://doi.org/10.3390/electronics11152457

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop