Next Article in Journal
Enhanced Digital Image Correlation Analysis of Ruptures with Enforced Traction Continuity Conditions Across Interfaces
Next Article in Special Issue
Computer-Aided Detection of Hyperacute Stroke Based on Relative Radiomic Patterns in Computed Tomography
Previous Article in Journal
Improving RNN Performance by Modelling Informative Missingness with Combined Indicators
Previous Article in Special Issue
A Trimmed Clustering-Based l1-Principal Component Analysis Model for Image Classification and Clustering Problems with Outliers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pedestrian Flow Tracking and Statistics of Monocular Camera Based on Convolutional Neural Network and Kalman Filter

1
Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
2
Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Science, Shenyang 110016, China
5
The Key Lab of Image Understanding and Computer Vision, Shenyang 110016, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(8), 1624; https://doi.org/10.3390/app9081624
Submission received: 19 March 2019 / Revised: 11 April 2019 / Accepted: 15 April 2019 / Published: 18 April 2019
(This article belongs to the Special Issue Advanced Intelligent Imaging Technology)

Abstract

:
Pedestrian flow statistics and analysis in public places is an important means to ensure urban safety. However, in recent years, a video-based pedestrian flow statistics algorithm mainly relies on binocular vision or a vertical downward camera, which has serious limitations on the application scene and counting area, and cannot make use of the large number of monocular cameras in the city. To solve this problem, we propose a pedestrian flow statistics algorithm based on monocular camera. Firstly, a convolution neural network is used to detect the pedestrian targets. Then, with a Kalman filter, the motion models for the targets are established. Based on these motion models, data association algorithm completes target tracking. Finally, the pedestrian flow is counted by the pedestrian counting method based on virtual blocks. The algorithm is tested on real scenes and public data sets. The experimental results show that the algorithm has high accuracy and strong real-time performance, which verifies the reliability of the algorithm.

1. Introduction

Pedestrian flow statistics is an important application in the field of computer vision [1]. It is a key technology in intelligent cities, intelligent retail, public place security and many other fields [2]. In recent years, with the continuous progress of intelligent city, pedestrian flow statistics has attracted more and more researchers and companies to participate in, and developed more and more statistics algorithms [3,4].
The earliest methods for pedestrian traffic statistics depend on manual statistics or bill statistics, which are either costly or have an impact on pedestrians. Then pedestrian flow statistics methods based on pressure sensor or photoelectric sensor are proposed. However, the accuracy of these methods are not good enough for dense pedestrian flow with severe occlusion. Due to the development of computer vision, pedestrian traffic statistics methods based on vertical downward stereo camera have emerged [5]. This kind of method is the most popular pedestrian traffic statistics method at present. Because the vertical downward camera can effectively avoid pedestrian occlusion, and the binocular-vision-based three-dimensional reconstruction algorithm can well filter out complex color background, the method has high accuracy, but the installation location and visual field are severely limited. It can only be applied to narrow indoor entrances, but not to commercial streets with wide outdoor entrances. At the same time, it is impossible to use the large number of front-down monocular cameras in the city.
Compared with vertical downward stereo camera, front-down monocular camera has wider vision, and because it can see the positive human face, it can better take into account security, criminal investigation and abnormal behavior early warning tasks. In front-down cameras, the image plane is about 45 degrees from the ground. The occlusion between the targets in front-down cameras is not as serious as that in front cameras. At the same time, the scale difference between targets is relatively small. There is a negative correlation between the size of the pedestrian target and the placement height of the camera.
However, due to the challenges of frequent occlusions, illumination changes, target scale changes, different fog concentrations in different distance and so on, pedestrian flow statistics algorithm based on front-down monocular camera puts forward higher requirements for detection and tracking algorithm. Thanks to the rapid progress of deep learning algorithm [6,7,8,9,10,11], the accuracy of pedestrian detection is constantly improving. More accurate detection results make the performance of tracking-by-detection method reach a higher level [12], which makes it possible to develop a high-precision pedestrian flow statistics algorithm based on front-down monocular camera.
In this paper, we propose a pedestrian flow tracking and statistics method based on front-down monocular camera. Firstly, we use convolutional neural network to detect pedestrians appearing in the camera, and modify the detection results based on intersection over union and aspect ratio. Secondly, we use Kalman filter to build uniform linear motion models for the detected pedestrians. Then pedestrian tracking is accomplished by data association algorithm. Finally, the virtual block method is used to count the target. We test the proposed algorithm using the real scenic spot entrance surveillance video, the F1 score of pedestrian flow statistics has reached 95%. At the same time, we compare the proposed multi-target tracking algorithm with other multi-target tracking algorithms on the 2DMOT2015 dataset. The MOTA of the proposed algorithm reaches 48.1, and the algorithm has advantages in computing speed.
The rest of this paper is organized as follows. Section 2 reviews the related work. The proposed algorithm is described in detail in Section 3. In Section 4, experiments and comparisons are carried out. Conclusion and analysis are presented in Section 5.

2. Related Work

The pedestrian flow statistics algorithm mainly includes three steps: pedestrian detection, multi-target tracking and pedestrian counting.
Current fast pedestrian detection methods are mainly divided into two categories. The first is pedestrian detection method based on background modeling, which mainly relies on the background modeling method to extract the foreground moving object, and uses classifier to judge whether the moving object is a pedestrian. GMM algorithm [13] and vibe algorithm [14] are the most representative. This kind of algorithm is fast in speed, but it cannot cope with the change of illumination and the jitter of camera very well. At the same time, it is difficult to distinguish dense objects or objects that are occluded from each other. The second kind of pedestrian detection algorithm is based on statistical learning. The algorithm has high accuracy and can cope with occlusion and environmental changes to a certain extent. The HOG + SVM [15] proposed by Dalal et al. is the most classical algorithm of this kind. Because of the popularity of deep learning methods in recent years, the accuracy of pedestrian detection algorithm based on convolutional neural network has reached an unprecedented height [16]. Detection methods based on deep learning can be divided into two-stage method and one-stage method. Faster R-CNN proposed by Shaoqing Ren [17] and its subsequent variants [18,19] belong to the two-stage method, which has advantages in detection and positioning accuracy. WeiLiu et al.’s SSD detection algorithm [20] and its subsequent variants [21,22] belong to one-stage algorithm, which achieves good accuracy and better real-time performance.
Multi-objective tracking methods can be divided into DBT (detection-based-tracking) methods and DFT (detection-free-tracking) methods based on initialization method. DFT methods needs to label the targets manually, and then track them in subsequent frames [23,24]. DBT methods completes tracking by detecting the targets in each frame and putting the targets into the tracklets [25,26,27]. DBT methods are more suitable for pedestrian traffic statistics applications because of the frequent appearance of new targets and the frequent disappearance of old ones. At the same time, according to the processing mode, multi-target tracking methods can be divided into online and offline algorithms. Online algorithm only uses the current frame of image sequence and several previous frames [28,29], which is more suitable for applications that need real-time implementation. The offline algorithm needs to use some future frames in the image sequence [30,31], which is more suitable for post-analysis and processing of video. Appearance model is widely used in the field of multi-target tracking. The appearance model has an important role in associating tracklets and detections. With the help of the appearance model, the ID scitches can be effectively suppressed. Ullah, M. et al. proposed a multi-target tracking method establishing appearance model with HoG descriptor [32]. Bae, S.H. et al. proposed a deep appearance learning method to learn a discriminative appearance model which can distinguish multiple objects with large appearance variations [33]. At the same time, the motion model has the same important role as the appearance model. Since the motion of the target in the image is usually relatively flat, the estimation of the trend of the target motion can predict the position of the target in the next frame, thus reducing the search area and even directly obtaining the tracking results.
The pedestrian counting method was originally road marking method. In this method, a mark is set on the road surface, and when the mark is covered, a pedestrian is judged to pass. Then Kryjak et al. proposed a counting method based on virtual lines [34]. The main idea is that when the target center passes through the virtual line, a pedestrian is judged to pass through. However, if there is a target hovering near the virtual line, it will seriously affect the counting accuracy. Later Xu et al. proposed a counting method based on double virtual lines [2]. In this method, two virtual lines are delineated, and the sequence of pedestrians passing through the virtual lines is judged to realize the counting.

3. Methodology

The proposed algorithm can be divided into three parts: pedestrian detection, multi-pedestrian tracking and pedestrian counting. The overall flow of the proposed method is shown in Figure 1.

3.1. Pedestrian Detection

Pedestrian detection is the first step of pedestrian flow statistics algorithm. The algorithm mainly improves from the yolov3 detection network [35]. In order to reduce the computational complexity of the algorithm, the darkent-53 network in the front of the network is replaced by a pruned and compressed VGG network [36], which reduces the computational complexity of the network from 65 Bflops to 39 Bflops. Because the target is pedestrian traffic statistics, too small anchor settings have no effect on improving detection accuracy, so kmeans algorithm is used to re-cluster the size of network anchor. Due to the large scale of the targets in the front-down camera, the last FPN structure of the yolov3 network is removed to reduce the computational complexity. The final computational complexity of the neural network is 34 Bflops. At the same time, the mAP (mean Average Precision) of the detector only reduce 1.61, from 74.49 to 72.88. The final network structure of the algorithm is shown in Figure 2.
In pedestrian flow statistics, pedestrian targets are very dense, and there is serious occlusion between the targets. The traditional non-maximum suppression (NMS) method can cause a part of the correct detection to be lost while removing redundant detection bounding boxes. So soft-NMS method is used to improve the non-maximum suppression [37]. Unlike the original non-maximum suppression method, as shown in Formula (1), soft-mns method does not directly remove the bounding box whose IOU exceeds the threshold and confidence is lower, but reduces the confidence of the detection box, which makes it more difficult for the correct detection to be removed incorrectly due to the dense targets. In Formula (1), d i is a detection result with score s i , d m is another detection result which has higher score than d i , N i represent the threshold of soft-NMS. As shown in Figure 3, after using soft-NMS, not only the redundant detection results can be correctly removed, but also the missing rate can be reduced.
s i = s i , I O U ( d m , d i ) < N i s i ( 1 I O U ( d m , d i ) , I O U ( d m , d i ) N i
In the front-down cameras, the objects close to the cameras can easily occlude the lower half of the objects farther from the cameras. On this basis, because of the lateral movement of the target, the bounding box of the farther target will change dramatically in height, which is not conducive to the following tracking operation. However, the top position and width of the bounding box will not be affected in such case. Based on this observation, we adjust the shape of bounding boxes by their width and the top position. For the bounding boxes whose aspect ratio are greater than 1:2.5, the heights of them are increased on the basis of fixing the top position and width, so that the aspect ratios are adjusted to 1:2.5. As shown in Figure 4, this scheme effectively reduces the deformation of the bounding box due to occlusion. At the same time, this method improves the positioning accuracy of the real center position of the occluded target, and is more conducive to the final counting task.

3.2. Multi- Pedestrian Tracking

The tracking algorithm is a multi-target tracking method based on detection results. The tracking method can be divided into two parts. Firstly, a motion model is built for the detected target, and new tracklets are built for new targets. The second part is the data association algorithm, which matches the target detected in each frame with the existing tracklets by the cost function to achieve the purpose of detection. The overall flow of the tracking algorithm is shown in Figure 5.
In the tracking algorithm, the linear motion model is chosen as the motion model of the tracklet. The model is based on Kalman filtering algorithm, and the target state is expressed as [ u , v , s , r , u , v , s ] T . Where ( u , v ) is the coordinate of the center position of the bounding box, and ( s , r ) are the scale and aspect ratio of the bounding box respectively. ( u , v ) is the speed of the target in horizontal and vertical directions, s represents the changing rate of the scale of the target. Since the aspect ratio of the bounding box is adjusted before, it is assumed that the aspect ratio of the target does not change here.
Data association algorithm uses Hungarian algorithm to match existing tracklets and detection results in current frame. The cost function is divided into three parts: IOU limit, scale changing limit and standardized distance. It is required that the IOU of the tracklet’s state and the detection result is larger than a certain threshold, and the scale change is lower than a certain threshold, otherwise the tracklet and the detection result will not match each other. Scale changes are described by the following formula:
D s c a l e = m i n ( m a x ( w 1 , w 2 ) m i n ( w 1 , w 2 ) , m a x ( h 1 , h 2 ) m i n ( h 1 , h 2 ) ) 1
where w 1 and h 1 are the width and height of the detection result respectively, w 2 and h 2 are the width and height of the tracklet state respectively.
On the basis of these two limitations, the matching degree between a detection result and a tracklet is mainly described by standardized distance. The standardized distance standardizes the pixel distance between the detector and the tracklet by the minimum width and height of them, which can effectively reduce the difference caused by the depth of field between the pixel distance and the actual distance. The standardized distance is shown as follows:
D s p = ( x 1 x 2 m i n ( w 1 , w 2 ) ) 2 + ( y 1 y 2 m i n ( h 1 , h 2 ) ) 2 )
where ( x 1 , y 1 ) is the coordinate of the detection result, ( x 2 , y 2 ) is the coordinate of the tracklet state.
For the successfully matched tracklets and detections, the status of the tracklets are updated by the positions of the detections. In this case, the state of the tracklet, including position, scale, velocity and scale change rate, is the optimal estimate obtained by Kalman filter. Unmatched detectors are candidates for new targets and candidate tracklets for them are established. If these candidate tracklets match detection results in consecutive multiple frames, they will be used as new tracklets for targets newly appear in these frames. Motion models established by detection results in the first frame of the image sequence are used as new tracklets immediately for the initialization of the sequence. The unmatched tracklet outputs the predicted results directly. In this case, the tracking state is completely determined by the prediction matrix, and the predicted tracklet moves in a straight line with a uniform speed decided by the state variables. This method can reduce the influence caused by occlusion or detector failure in a short time. When a tracklet fails to match any detection result in continuous multiple frames, it is considered that the target tracked by the tracklet has disappeared and the tracklet is deleted.

3.3. Pedestrian Counting

A pedestrian counting algorithm based on virtual blocks is proposed. Similar to the counting method based on double virtual lines, the pedestrian counting algorithm based on virtual blocks counts the number of pedestrians according to the sequence of blocks passed by the pedestrian detection center. This method shortens the time requirement of continuous target tracking and makes the holistic algorithm more robust to occlusion.
Block-based counting algorithms need to delimit the beginning area, count area and end area, as shown in Figure 6. If the center of the target is initialized in the beginning area and reaches the end area after passing through the counting area, the count is made once. Two-way counting can be realized by delimiting regions in different order.
Compared with the counting method based on double virtual lines, the block-based method can better adapt to the counting area with different shapes, which is more suitable for the application scenarios of front-down cameras. In addition, the block-based method is easier to achieve the effect of counting part of the road area through flexible setting of counting area. In addition, by setting the start and end areas, it is easier to count only for the target entering from a specific entry or leaving from a specific exit.

4. Experiment

4.1. The Performance of Pedestrian Flow Statistics Algorithms in Real Scene

Since there is no dataset specifically for pedestrian flow statistics, we use the image sequence captured by front-down cameras at the entrance of crowded scenic spots to verify the detection and counting effect of the proposed algorithm.
To develop an algorithm with high robustness against the changes of illumination and fog concentration, we have manually detected and labeled the images captured by the camera, and made a new dataset. As a detection dataset, five half-hour video recordings were collected. There were intense changes in illumination in the videos, and some videos have dense fog. The frame rate of video recordings is 25 fps, and an image is saved every 12 frames. The total number of annotated images is 18,750, including 282,674 pedestrian bounding boxes. Among them, 2000 images from the end of two videos with different perspectives are used as the test set, while the remaining 16,750 images are used as the training set. After training, the detection effect of the algorithm is evaluated, and the most popular evaluation indicators in the detection field, such as precision, recall, mAP (mean Average Precision), are selected as the evaluation indicators. The detection performance is shown in Table 1. Compared with detectors trained only with coco datasets, the robustness of the detector trained with new images to the changes of illumination and fog concentration is greatly improved.
For pedestrian counting effect, we selected five other videos to evaluate. By observing the counting results manually, the missing alarm and false alarm of the counting algorithm are counted. Finally, the counting algorithm is evaluated by the precision, recall and F1-Score of the counting algorithm. The evaluation results of the algorithm in five videos are shown in Table 2.
It can be seen that the mean F1-score of the algorithm is over 95% in the five videos, which proves the effectiveness of the algorithm. The running speed of the algorithm on NVIDIA GTX1060 GPU is 28.4 FPS, which has good real-time performance. The actual running effect is shown in Figure 7.

4.2. The Performance of Tracking-by-Detection Algorithm Compared with Other Algorithms

In pedestrian flow statistics algorithms, the performance of detecting and tracking algorithms has an important impact on the counting results. Therefore, we evaluate the proposed algorithm by comparing the comprehensive performance of the proposed detecting and tracking algorithm with other algorithms.
Here we choose to use the 2DMOT2015 benchmarks [38] to evaluate the performance of the algorithm quantitatively. 2DMOT2015 benchmark is a well-known framework for the fair evaluation of multi-pedestrian tracking algorithms. The dataset includes 22 image sequences, of which 11 are training sets and 11 are test sets. Image sequences come from several influential pedestrian detection and tracking datasets, including KITTI, ADL, ETH, PETS and TUD.
The main evaluation indicators are as in Table 3.
Among them, MOTA is a comprehensive evaluation of FP, FN and IDSW. Its formula is as follows:
MOTA = 1 t ( FN t + FP t + IDSW t ) t G T t
where G T t is the number of ground truth in frame t.
MOTP focuses on the average difference between TP and its corresponding ground truth. The formula is as follows:
MOTP = t , i d t , i t c t
where d t , i denotes the overlap of the bounding box i with its corresponding ground truth, c t denotes the amount of bounding boxes which match ground truths successfully.
To verify the comprehensive performance of detector and tracker, we compare the proposed algorithm with other algorithms using private detector, including MDP_Subcnn [39], DMT [40] and Sort [41]. The specific comparison results are shown in Table 4.
It can be seen that the proposed algorithm has high accuracy. The actual effect of the algorithm on the MOT2015 dataset is shown in Figure 8.

5. Conclusions

In this paper, we propose a pedestrian flow tracking and statistics algorithm for front-down monocular camera. The algorithm relies on convolutional neural network for real-time pedestrian detection, and uses Kalman filter linear motion model and data association algorithm to track pedestrian targets. Finally, a counting method based on virtual blocks is proposed to complete pedestrian flow statistics. We use real scene videos to evaluate the counting performance of the algorithm. At the same time, we compare the detection and tracking performance of the algorithm with other algorithms using a public dataset, MOT 2015, which proves the effectiveness of the algorithm. The experiment results show that the algorithm has good accuracy and real-time performance, and has high application value. Although the algorithm has achieved good results, there are still some shortcomings, which need further improvement. Since the accuracy of the algorithm will be affected when the pedestrian is seriously occluded, future work is to further improve the tracking accuracy of the algorithm in the case of serious occlusion by means of optical flow and re-identifying appearance model.

Author Contributions

Conceptualization, M.H.; Methodology, M.H.; Software, M.H.; Validation, M.H., B.H. and Z.C.; Formal Analysis, M.H.; Investigation, M.H.; Resources, B.H. and Z.C.; Data Curation, M.H.; Writing–Original Draft Preparation, M.H.; Writing–Review and Editing, H.L.; Visualization, M.H.; Supervision, H.L.; Project Administration, H.L.; Funding Acquisition, H.L.

Funding

This research was funded by Hainan Heaven Reward Security Technology Co. Ltd.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Del Pizzo, L.; Foggia, P.; Greco, A.; Percannella, G.; Vento, M. Counting people by RGB or depth overhead cameras. Pattern Recognit. Lett. 2016, 81, 41–50. [Google Scholar] [CrossRef]
  2. Coşkun, A.; Kara, A.; Parlaktuna, M.; Ozkan, M.; Parlaktuna, O. People counting system by using kinect sensor. In Proceedings of the 2015 IEEE International Symposium on Innovations in Intelligent SysTems and Applications (INISTA), Taipei, Taiwan, 2–4 September 2015; pp. 1–7. [Google Scholar]
  3. Verma, N.K.; Dev, R.; Maurya, S.; Dhar, N.K.; Agrawal, P. People Counting with Overhead Camera Using Fuzzy-Based Detector. In Computational Intelligence: Theories, Applications and Future Directions—Volume I; Springer: Berlin, Germany, 2019; pp. 589–601. [Google Scholar]
  4. Kopaczewski, K.; Szczodrak, M.; Czyzewski, A.; Krawczyk, H. A method for counting people attending large public events. Multimed. Tools Appl. 2015, 74, 4289–4301. [Google Scholar] [CrossRef]
  5. Beymer, D. Person counting using stereo. In Proceedings of the Workshop on Human Motion, Austin, TX, USA, 7–8 December 2000; pp. 127–133. [Google Scholar]
  6. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  8. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  9. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference On Computer Vision And Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference On Computer Vision And Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  11. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  12. Bochinski, E.; Eiselein, V.; Sikora, T. High-speed tracking-by-detection without using image information. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar]
  13. Stauffer, C.; Grimson, W.E.L. Adaptive background mixture models for real-time tracking. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), Fort Collins, CO, USA, 23–25 June 1999; p. 2246. [Google Scholar]
  14. Barnich, O.; Van Droogenbroeck, M. ViBe: A powerful random technique to estimate the background in video sequences. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP 2009, Taipei, Taiwan, 19–24 April 2009; pp. 945–948. [Google Scholar]
  15. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR 2005, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  16. Zhou, C.; Yuan, J. Bi-box Regression for Pedestrian Detection and Occlusion Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 135–151. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 6, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  18. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  19. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; Volume 1, p. 4. [Google Scholar]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference On Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Berlin, Germany, 2016; pp. 21–37. [Google Scholar]
  21. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. DSSD: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  22. Li, Z.; Zhou, F. FSSD: Feature Fusion Single Shot Multibox Detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  23. Hu, W.; Li, X.; Luo, W.; Zhang, X.; Maybank, S.; Zhang, Z. Single and multiple object tracking using log-Euclidean Riemannian subspace and block-division appearance model. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2420–2440. [Google Scholar] [PubMed]
  24. Zhang, L.; van der Maaten, L. Structure preserving object tracking. In Proceedings of the IEEE Conference On Computer Vision And Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1838–1845. [Google Scholar]
  25. Berclaz, J.; Fleuret, F.; Turetken, E.; Fua, P. Multiple object tracking using k-shortest paths optimization. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1806–1819. [Google Scholar] [CrossRef] [PubMed]
  26. Breitenstein, M.D.; Reichlin, F.; Leibe, B.; Koller-Meier, E.; Van Gool, L. Robust tracking-by-detection using a detector confidence particle filter. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 1515–1522. [Google Scholar]
  27. Ess, A.; Leibe, B.; Schindler, K.; Van Gool, L. Robust multiperson tracking from a mobile platform. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 1831–1846. [Google Scholar] [CrossRef] [PubMed]
  28. Choi, W.; Pantofaru, C.; Savarese, S. A general framework for tracking multiple people from a moving camera. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1577–1591. [Google Scholar] [CrossRef] [PubMed]
  29. Khan, Z.; Balch, T.; Dellaert, F. MCMC-based particle filtering for tracking a variable number of interacting targets. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1805–1819. [Google Scholar] [CrossRef] [PubMed]
  30. Kuo, C.H.; Huang, C.; Nevatia, R. Multi-target tracking by on-line learned discriminative appearance models. In Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 685–692. [Google Scholar]
  31. Milan, A.; Roth, S.; Schindler, K. Continuous energy minimization for multitarget tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 58–72. [Google Scholar] [CrossRef] [PubMed]
  32. Ullah, M.; Cheikh, F.A.; Imran, A.S. Hog based real-time multi-target tracking in bayesian framework. In Proceedings of the 2016 13th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Colorado Springs, CO, USA, 23–26 August 2016; pp. 416–422. [Google Scholar]
  33. Bae, S.H.; Yoon, K.J. Confidence-based data association and discriminative deep appearance learning for robust online multi-object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 595–610. [Google Scholar] [CrossRef] [PubMed]
  34. Kryjak, T.; Komorkiewicz, M.; Gorgon, M. Hardware-software implementation of vehicle detection and counting using virtual detection lines. In Proceedings of the 2014 Conference on IEEE Design and Architectures for Signal and Image Processing (DASIP), Madrid, Spain, 8–10 October 2014; pp. 1–8. [Google Scholar]
  35. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  36. He, Y.; Zhang, X.; Sun, J. Channel pruning for accelerating very deep neural networks. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; Volume 2. [Google Scholar]
  37. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-nms—Improving object detection with one line of code. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5562–5570. [Google Scholar]
  38. Leal-Taixé, L.; Milan, A.; Reid, I.; Roth, S.; Schindler, K. Motchallenge 2015: Towards a benchmark for multi-target tracking. arXiv 2015, arXiv:1504.01942. [Google Scholar]
  39. Xiang, Y.; Alahi, A.; Savarese, S. Learning to track: Online multi-object tracking by decision making. In Proceedings of the IEEE International Conference On Computer Vision, RegióN Metropolitana, Chile, 11–18 December 2015; pp. 4705–4713. [Google Scholar]
  40. Kim, H.U.; Kim, C.S. CDT: Cooperative detection and tracking for tracing multiple objects in video sequences. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Berlin, Germany, 2016; pp. 851–867. [Google Scholar]
  41. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on. IEEE Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
Figure 1. Overall flow chart of the proposed algorithm.
Figure 1. Overall flow chart of the proposed algorithm.
Applsci 09 01624 g001
Figure 2. Network structure of detection algorithms. The blue layers are convolutional layers, the red layers are max-pooling layers, and the yellow layers are detection layers.
Figure 2. Network structure of detection algorithms. The blue layers are convolutional layers, the red layers are max-pooling layers, and the yellow layers are detection layers.
Applsci 09 01624 g002
Figure 3. Detection result with soft-NMS (left) and NMS (right). The bounding box for the person in yellow is not miss deleted as redundant detection with soft-NMS.
Figure 3. Detection result with soft-NMS (left) and NMS (right). The bounding box for the person in yellow is not miss deleted as redundant detection with soft-NMS.
Applsci 09 01624 g003
Figure 4. The deformation of the bounding box of a single target without (left) and with (right) shape adjusting.
Figure 4. The deformation of the bounding box of a single target without (left) and with (right) shape adjusting.
Applsci 09 01624 g004
Figure 5. Overall flow of the tracking algorithm.
Figure 5. Overall flow of the tracking algorithm.
Applsci 09 01624 g005
Figure 6. Virtual block-based method, pedestrian counting method.
Figure 6. Virtual block-based method, pedestrian counting method.
Applsci 09 01624 g006
Figure 7. The actual operation effect of the algorithm.
Figure 7. The actual operation effect of the algorithm.
Applsci 09 01624 g007
Figure 8. Tracking result on 2DMOT15 dataset.
Figure 8. Tracking result on 2DMOT15 dataset.
Applsci 09 01624 g008
Table 1. Detection performance of detector trained with different data.
Table 1. Detection performance of detector trained with different data.
Precision↑Recall↑F1-Score↑TP↑FP↓FN↓mAP↑
detector (coco)0.870.670.76177862644888772.88%
detector (coco + new data)0.860.790.82211563550551783.39%
Table 2. The evaluation results of the algorithm in test videos.
Table 2. The evaluation results of the algorithm in test videos.
SequenceCount ResultGTFPFNTPPrecisionRecallF1-Score
Seq1552567213653196.20%93.65%94.91%
Seq2411412181939395.62%95.39%95.50%
Seq3571572212255096.32%96.15%96.24%
Seq4923955185090598.05%94.76%96.38%
Seq5669693164065397.61%94.23%95.89%
TOTAL3126319994167303296.99%94.78%95.87%
Table 3. Main evaluation indicators.
Table 3. Main evaluation indicators.
MeasureBetterDescription
MOTAhigherMultiple Object Tracking Accuracy.
MOTPhigherMultiple Object Tracking Precision.
IDF1higherThe ratio of correctly identified detections.
MThigherMostly tracked targets.
MLlowerMostly lost targets.
FPlowerThe total number of false positives.
FNlowerThe total number of false negatives.
ID Sw.lowerThe total number of identity switches.
Table 4. The performance of proposed tracking-by-detection algorithm compared with other algorithms in 2DMOT2015 dataset. The green and yellow colors indicate the best and the second best algorithm in each measure.
Table 4. The performance of proposed tracking-by-detection algorithm compared with other algorithms in 2DMOT2015 dataset. The green and yellow colors indicate the best and the second best algorithm in each measure.
SequenceAlgorithmMOTAIDF1MOTPMTMLFPFNID Sw
TUD-CrossingMDP_Subcnn78.974.576.70.6920321956
DMT7056.373.30.61507322929
Sort67.557.674.50.4620.0773230818
PFS(our)74.649.976.20.61504919338
PETS09-S2L2MDP_Subcnn47.536.872.60.0710.0953414524196
DMT47.736.870.40.2140.195024429113
Sort27.423.667.400.2628065958240
PFS(our)56.73274.40.190.0483583519295
ETH-JelmoliMDP_Subcnn48.265.777.30.3560.2224928149
DMT48.262.8750.60.11175852926
Sort3952.974.10.20.289439107138
PFS(our)61.55976.20.3560.26717675446
ETH-LinthescherMDP_Subcnn63.967.177.10.2440.31495265770
DMT60.560.276.40.4370.24414251963138
Sort52.254.573.80.1470.4063973725144
PFS(our)62.857.576.80.3350.2237352374215
ETH-CrossingMDP_Subcnn63.876.579.50.1920.269642936
DMT5954.781.50.3080.30816922121
Sort55.449.780.30.1540.3855836821
PFS(our)62.453.979.50.3080.07713521527
AVG-TownCentreMDP_Subcnn49.564.570.10.3890.15513812106121
DMT45.554.668.50.2830.2746533127117
Sort27.245.167.40.0580.27911113930162
PFS(our)47.559.769.50.3720.11514832094178
ADL-Rundle-1MDP_Subcnn33.449.972.40.34402899323070
DMT2642.270.10.2190.2811697514647
Sort20.331.572.50.1880.37514935812108
PFS(our)39.838.673.80.250.09414443998164
ADL-Rundle-3MDP_Subcnn44.951.679.60.2050.159793475256
DMT43.345.475.40.2950.1361168451784
Sort37.443.4770.2950.1821498476599
PFS(our)34.834.1800.2730.15920124422197
KITTI-16MDP_Subcnn5066.670.30.3530.05926256622
DMT44.760.569.30.235023269019
Sort34.642.870.10.1180.05914493830
PFS(our)5056.773.20.2350.05926854439
KITTI-19MDP_Subcnn40.961.868.10.2420.0811143196551
DMT45.755.572.80.3060.097884194672
Sort29.146.468.400.258855285279
PFS(our)37.846.4690.2740.12913271890109
Venice-1MDP_Subcnn42.648.4760.3530.235729186721
DMT32.43870.90.2940.412527253818
Sort24.724.469.70.1180.471485288862
PFS(our)43.740.175.80.2940.176435207164
TOTALMDP_Subcnn47.555.774.20.30.186863122,969628
DMT44.549.272.90.3470.221808825,335684
Sort33.440.472.10.1170.309731832,6151001
PFS(our)48.14574.70.3270.15842222,0741372

Share and Cite

MDPI and ACS Style

He, M.; Luo, H.; Hui, B.; Chang, Z. Pedestrian Flow Tracking and Statistics of Monocular Camera Based on Convolutional Neural Network and Kalman Filter. Appl. Sci. 2019, 9, 1624. https://doi.org/10.3390/app9081624

AMA Style

He M, Luo H, Hui B, Chang Z. Pedestrian Flow Tracking and Statistics of Monocular Camera Based on Convolutional Neural Network and Kalman Filter. Applied Sciences. 2019; 9(8):1624. https://doi.org/10.3390/app9081624

Chicago/Turabian Style

He, Miao, Haibo Luo, Bin Hui, and Zheng Chang. 2019. "Pedestrian Flow Tracking and Statistics of Monocular Camera Based on Convolutional Neural Network and Kalman Filter" Applied Sciences 9, no. 8: 1624. https://doi.org/10.3390/app9081624

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop