Next Article in Journal
Assessing the Global Cognition of Community-Dwelling Older Adults Using Motor and Sensory Factors: A Cross-Sectional Feasibility Study
Next Article in Special Issue
A Generative Approach to Person Reidentification
Previous Article in Journal
Human Factors Considerations for Quantifiable Human States in Physical Human–Robot Interaction: A Literature Review
Previous Article in Special Issue
Discriminatively Unsupervised Learning Person Re-Identification via Considering Complicated Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Approaches to Improve the Quality of Person Re-Identification for Practical Use

1
Faculty of Computational Mathematics and Cybernetics, Moscow State University, 119991 Moscow, Russia
2
Video Analysis Technologies LLC, 119634 Moscow, Russia
3
Faculty of Computer Science, National Research University Higher School of Economics, 109028 Moscow, Russia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(17), 7382; https://doi.org/10.3390/s23177382
Submission received: 8 July 2023 / Revised: 13 August 2023 / Accepted: 22 August 2023 / Published: 24 August 2023
(This article belongs to the Special Issue Person Re-Identification Based on Computer Vision)

Abstract

:
The idea of the person re-identification (Re-ID) task is to find the person depicted in the query image among other images obtained from different cameras. Algorithms solving this task have important practical applications, such as illegal action prevention and searching for missing persons through a smart city’s video surveillance. In most of the papers devoted to the problem under consideration, the authors propose complex algorithms to achieve a better quality of person Re-ID. Some of these methods cannot be used in practice due to technical limitations. In this paper, we propose several approaches that can be used in almost all popular modern re-identification algorithms to improve the quality of the problem being solved and do not practically increase the computational complexity of algorithms. In real-world data, bad images can be fed into the input of the Re-ID algorithm; therefore, the new Filter Module is proposed in this paper, designed to pre-filter input data before feeding the data to the main re-identification algorithm. The Filter Module improves the quality of the baseline by 2.6 % according to the R a n k 1 metric and 3.4 % according to the m A P metric on the Market-1501 dataset. Furthermore, in this paper, a fully automated data collection strategy from surveillance cameras for self-supervised pre-training is proposed in order to increase the generality of neural networks on real-world data. The use of self-supervised pre-training on the data collected using the proposed strategy improves the quality of cross-domain upper-body Re-ID on the DukeMTMC-reID dataset by 1.0 % according to the R a n k 1 and m A P metrics.

1. Introduction

The task of person re-identification (Re-ID) is to identify the same person in different images obtained from different cameras or at different points in time. That is, a set of images is given, which depict different people in different conditions and from different angles. The task is to determine whether two different images refer to the same person or not. This task belongs to metric learning tasks.
This problem arises in the context of video surveillance and security systems, where it is often necessary to track the movement of people in different zones or places using different video cameras. For example, re-identification can be used at airports, malls, stadiums, or other public places where there are many people and few cameras. The task of person Re-ID is practically important for several reasons:
  • Crime Control. Re-identification algorithms help in identifying criminals and assist investigators in investigating crimes, including the search for missing persons, abductions, and other serious crimes;
  • Improving the retail customer experience. In retail, person Re-ID algorithms can help to improve the customer experience by providing personalized offers or services using the analysis of a person’s movement around the store;
  • Analysis of sports events. Re-identification algorithms are able to automate the processes associated with the analysis of the actions of athletes (for example, the analysis of the movement of football players across the field).
The effectiveness of solving the problem under consideration is of great importance for improving security, ensuring public order and optimizing video surveillance systems. In this regard, research in this area continues, and new techniques and algorithms are constantly being developed for more accurate person Re-ID. Since the algorithms for solving the problem under consideration are used in practice, it is also important to take into account the computational complexity of the re-identification methods during research.
Person Re-ID implies working with real-world data; therefore, due to the variety of situations, when solving this task, we have to face a lot of nuances. Conditionally, there are three main problems that arise when solving this task:
  • Lots of object occlusions for re-identification: Due to occlusion, the area for re-identification is reduced, and occlusions can make noise, for example, parts of other objects can get into the image;
  • Domain gap problem: The domain in the training data is usually different from the real domain in which the algorithm of person Re-ID will work;
  • A small amount of data suitable for training and testing re-identification algorithms: Usually, person Re-ID requires data captured in a multi-camera scenario. Such data are difficult to collect and also expensive to mark up.
One obvious approach to tackle these challenges is by training re-identification algorithms using extensive datasets that have the potential to encompass a wide range of real-world scenarios. However, implementing this approach comes with various aspects that render it either practically infeasible or excessively expensive. Firstly, the datasets must exhibit significant diversity, capturing various situations and visual elements, such as lighting, quality, and angles. Secondly, achieving this diversity requires a substantial amount of data, involving several million images that need manual annotation. This process demands the participation of a large number of markers, resulting in considerable monetary costs. Thirdly, as mentioned earlier, collecting data for person Re-ID is highly problematic due to the unique nature of the task, particularly involving multi-camera scenarios.
Researchers have already made attempts to collect and mark up datasets for the problem under consideration [1,2,3,4]. However, as practice has shown, the data presented was not enough to completely solve the problems mentioned above. Therefore, researchers are studying other approaches that will solve the problems mentioned above.
One of the ways to solve the problem of object occlusions is to feed the algorithm, along with the input image, a binary mask that characterizes the position and shape of the human body that needs to be re-identified [5,6]. This approach contributes to a better perception of the information in the image by the neural network. The mask helps the neural network to focus on a particular person rather than on the background or an overlapping object. There are also solutions where semantic segmentation is used instead of the binary masks [7,8,9]. The essence of semantic segmentation is not in the allocation of the necessary object, but in the classification of the masked objects (for example, outerwear, shoes, etc.). This solution allows the neural network to search for a person more confidently, but this method has its drawbacks. To train the neural network to distinguish objects, it is necessary to manually mark them up in the training data, and this is a very difficult and expensive task.
In some works, the problems described above are proposed to be solved with the help of augmentations. For example, [10,11] uses a special Random Erasing strategy [12] as one of the augmentation approaches, the essence of which is as follows. Random figures are added to images of people in arbitrary places, thus simulating occlusions. This approach makes it possible to increase the amount of data for training and, at the same time, partially solves the problem of frequent occlusions.
There is also another approach to solving the problem with a lack of data—the use more advanced loss functions [13,14,15], which allows training the neural network on a smaller number of images, but at the same time maintaining (or even improving) the results. Loss functions play an important role in training the neural network to focus on a particular person. That is why, when using more advanced functions, less data may be required to train the neural network.
Some researchers suggest increasing training datasets at the expense of synthetic data generated either automatically or semi-automatically. Most often, data generation for training occurs using game engines [16] (for example, Unreal Engine) or directly from the games [17] (for example, Grand Theft Auto V). That is, scenes are generated with a predetermined scenario of the characters’ behavior, and marked-up data are collected in almost automatic mode, which can then be used to train re-identification algorithms. Through the use of software tools, researchers have an almost limitless possibility of generating training samples; with the help of game engines, we can set arbitrary behavior of characters, change their clothes, appearance, environment, etc.
In this paper, we propose our approaches to solving the problems mentioned above that arise in the task of person Re-ID. These approaches were the result of our research aimed at improving the quality of re-identification for practical use. In summary, the contributions of this paper are concluded as follows:
  • We will show that larger and advanced neural networks are able to solve the problem of person Re-ID better. However, since the task of re-identification is practically important, it is necessary to find a balance between the quality of the algorithm and its speed.
  • We will consider the problem of the algorithm’s stability to image changes and noise. In this paper, we propose using the Jensen–Shannon divergence and AugMix augmentation [18] to solve this problem.
  • In practice, there are frequent cases where incorrect images are fed to the input of the algorithm. Therefore, in this paper, we propose a simple Filter Module designed to pre-filter input data before feeding the data to the main re-identification algorithm.
  • We will show that due to occlusions, in practice, it will be better to use upper-body detections instead of full-body detections for person Re-ID.
  • In this paper, we will introduce a fully automated data collection strategy from surveillance cameras for self-supervised pre-training to solve the lack of data to train re-identification algorithms and increase the generality of neural networks for working with real-world data in practice.

2. Related Work

There are three directions in the task of person Re-ID: classical person re-identification, clothing-change person re-identification, and visible–infrared person re-identification. In this paper, classical person re-identification is considered. Despite the fact that the problems differ in conditions, the ideas of the algorithms solving them may be similar. Therefore, it is useful to study methods for solving all problems. Further, some approaches to solving these problems will be considered. Furthermore, in this paper, it is proposed to use self-supervised pre-training to improve the quality of person Re-ID (Section 3.5.2); therefore, in this section, some existing approaches for self-supervised pre-training will also be considered.

2.1. Classical Person Re-Identification

The classical person re-identification involves searching for a person on different cameras without taking into account the possible change in clothes. This area has received a special development with the growing popularity of convolutional neural networks [10,11,19,20,21,22]. Furthermore, recently, works related to the problem under consideration have begun to appear, in which transformer-based architectures are used [23,24,25,26]. Due to the fact that there are problems with data in the task of person Re-ID (Section 1), methods based on unsupervised learning have recently begun to develop. There are two types of unsupervised re-identification algorithms: unsupervised domain adaptation and fully unsupervised. Algorithms from the first class [27,28,29] attempt to transfer knowledge from existing labeled data to unlabeled target data. Fully unsupervised methods [30,31,32] imply model training, in which there is no markup at all.

2.2. Clothing-Change Person Re-Identification

The idea of the clothing-change person re-identification is to find the person depicted in the query image among other images obtained from different cameras, provided that people can change clothes at different times. There are approaches that use keypoints of the human body to solve this problem [33], as well as human contour sketches [34]. Among other things, there are works [35,36] in which knowledge about the shape of the body, extracted from human silhouettes, is used for person Re-ID. Furthermore, one of the distinguishing factors of each person is gait, so there are algorithms that try to re-identify people by using video recordings [37,38].

2.3. Visible-Infrared Person Re-Identification

Modern surveillance systems often operate in two modes: visible mode during the day (RGB images) and infrared mode at night (grayscale images). This creates an additional challenge beyond classical re-identification. An additional task involves matching the visible image of the target with the infrared image of the corresponding person. This cross-modal image matching task is called visible–infrared person Re-ID. In [39,40,41], two separate convolutional neural networks are employed as backbones for feature extraction from infrared and visible images. Based on the extracted features, the similarity between the two images is calculated, and then optimization occurs, during which the similarity increases for pairs with the same person and decreases otherwise. There are also papers [42,43] that propose augmentations that mix RGB and grayscale images.

2.4. Self-Supervised Pre-Training

The main idea of the self-supervised pre-training is to learn robust representations without expensive labels or annotations. Most self-supervised approaches [44,45,46] are based on the idea of contrastive learning, where two augmentations of the same image come closer to each other in space and move away from other images. Research has shown that for better self-supervised pre-training, a large number of negative examples are needed. In [47], the authors propose using a large batch for self-supervised pre-training to increase the number of negative samples. In another work [48], all embeddings of all images from the dataset are stored in the memory bank. However, this approach is bad from the point of view of memory consumption, so in [49,50], using a momentum encoder and a queue-like memory bank to dynamically update negative samples was proposed. In [51], authors propose using variance, invariance, and covariance regularization to learn robust representations.

3. Proposed Approaches

To demonstrate the effectiveness of the approaches proposed in this paper, we use a modified version of the neural network from [10] as a baseline (Figure 1). This baseline is simple to implement and is an ideal option for subsequent improvements. Despite its simplicity, the baseline has competitive performance in benchmarks for the task of person re-identification. Furthermore, it is important to note that all the proposed approaches can be applied to more advanced re-identification algorithms.

3.1. Using Larger and Advanced Neural Networks as the Encoder

This paper hypothesizes that the use of larger and more advanced neural networks as the encoder can improve the quality of person Re-ID. It can only be verified experimentally (Section 4.3.1); however, the use of larger neural networks can significantly increase the computational costs and the running time of the algorithm, but this is unacceptable for practical use. Therefore, a more difficult task is finding a tradeoff between the quality of the encoder and its performance.
In this paper, we propose replacing the ResNet50 [52], used in baseline as the encoder, with Res2Net50 [53] because:
  • Res2Net has an increased receptive field due to the replacement of one convolutional layer in the residual block in ResNet with a group of smaller ones that are interconnected with each other; moreover, an increased receptive field may favorably affect the quality of person Re-ID;
  • Res2Net shows higher accuracy in the classification problem on ImageNet [54] than ResNet;
  • The use of Res2Net as the encoder in baseline does not greatly affect the computational complexity of the algorithm. The time measurements presented in Table 1 confirm this.
Thus, Res2Net50 can improve the quality of solving the problem under consideration and at the same time slightly increase computational costs.

3.2. Improving the Stability of the Algorithm to Image Changes and Noise

In real-world scenarios, due to occlusions or detector mistakes, there may be cases where the detector finds only a part of the human body. The use of such detections for re-identification can seriously reduce the quality of person Re-ID. Therefore, current re-identification algorithms become unstable with image changes and noise because the slightest changes in images can seriously change the distribution of neural network outputs.
To solve this problem, in our previous work [57], we proposed the new Random Size Augmentation, which randomly changes the size of the area for re-identification during algorithm training. In this paper, we propose an improved approach to solving this problem—the use of AugMix augmentation, as well as JS Loss instead of Triplet Loss [58]. Together, the mentioned augmentation and loss function pursue the following idea: minimize the Jensen–Shannon divergence of the posterior distributions for the original image and its augmentations:
M = p o r i g + p a u g m i x 1 + p a u g m i x 2 3 ,
J S ( p o r i g , p a u g m i x 1 , p a u g m i x 2 ) = K L [ p o r i g | | M ] + K L [ p a u g m i x 1 | | M ] + K L [ p a u g m i x 2 | | M ] 3 ,
where p o r i g , p a u g m i x 1 , and p a u g m i x 2 are the neural network output distribution for the original image and its two AugMix augmentations, respectively. K L represents the Kullback–Leibler divergence.
Thus, the distribution of neural network outputs for modified images should become similar to the distribution of neural network outputs for the original image. This makes the algorithm less sensitive to image changes and noise and improves the quality of the solution of the problem under consideration. This is also confirmed by the experimental estimate given in Section 4.3.2.

3.3. Filter Module

In practice, as noted above in Section 3.2, detection errors and errors of other components of the video analytics system are possible. Therefore, there is a possibility that the re-identification algorithm will obtain a deliberately incorrect image (for example, an image without a person) at the input. Furthermore, in practice, most often for one person there are several images at once, so it is useful to evaluate the suitability of each of the images for subsequent re-identification (Figure 2).
In this paper, we propose a new simple Filter Module, which is a lightweight neural network binary classifier based on MobileNetV2 [59] (Figure 3). The Filter Module predicts the suitability of an image for subsequent re-identification. That is, at the inference stage, incorrect input images are filtered before they are fed to the neural network for Re-ID.
Thus, if only correct data are fed to the input of the re-identification algorithm, then it will be able to give only confident answers, which is a key point since most of the current Re-ID algorithms do not take into account the correctness of the input images and try to make a prediction even for those examples on which this is impossible to do, and it is undesirable both in the theoretical understanding of the problem and in the practical one.

3.4. Upper-Body Re-ID

Re-identification methods are being implemented in video surveillance systems, where large crowds of people are dealt with most often. Due to the large crowds, most of the human body, which is used for Re-ID, is not visible (Figure 4). Using full-body person re-identification in such a scenario can lead to poor performance of the algorithm due to occlusions and detector errors. Therefore, in this paper, we propose an alternative approach, in which the input to the Re-ID algorithm is not the detection of a full body (full-body Re-ID), as is customary in classical methods, but the detection of the upper body of a person (upper-body Re-ID).
The validity of this approach was confirmed by the corresponding experiments in our previous works [57,60] on the implementation of re-identification algorithms in tracking algorithms for estimating the waiting time in queues.

3.5. Fully Automated Data Collection Strategy and Self-Supervised Pre-Training

As mentioned in Section 1, the lack of sufficient data to train person re-identification algorithms is one of the main problems that researchers have to face when solving this task.
This paper considers an approach to improve the quality of re-identification through self-supervised pre-training. However, for high-quality pre-training of the neural network, it is necessary to have a large and diverse dataset of images of people. Therefore, in this paper, we propose a fully automated strategy for collecting such a dataset.

3.5.1. Fully Automated Data Collection Strategy

The proposed automated data collection strategy for self-supervised pre-training consists of the following steps:
  • Using the tracking algorithm [57], the tracks of the movement of people are built on video recordings collected from open sources;
  • Using the obtained tracks, images of people are cropped from video frames of video recordings;
  • After that, the crops of people are filtered automatically: false detections and cases when people are not moving during the whole video are removed, etc.
The suggested approach enables the collection of large volumes of data for neural network pre-training from open sources with minimal human involvement. By utilizing a tracking algorithm, the resulting dataset comprises multiple examples for each individual included in the dataset, which positively impacts the pre-training quality of the neural network for the person re-identification task. Furthermore, considering that each track corresponds to one person, automatic annotation can be obtained, facilitating self-supervised pre-training as well.
Section 4.1.3 provides information about which dataset for self-supervised pre-training was collected using the proposed strategy.

3.5.2. Self-Supervised Pre-Training

Most often, computer vision researchers working with neural networks do the following: a neural network pre-trained on a large ImageNet dataset for the classification task is taken, and this network is fine-tuned on target data (for example, images of people) to solve the final task.
In this paper, we hypothesize that if we somehow pre-train the neural network on data that are most similar to the target (in this case, on crops of people) rather than images of 1000 classes from ImageNet, then after fine-tuning this network on data for the final task, it is possible to achieve a better quality of its solution. This paper proposes achieving this through self-supervised pre-training.
Most self-supervised approaches are based on the idea of contrastive learning. The essence of contrastive learning lies in the fact that the neural network is trained to bring together positive examples (images of the same class) in space and move negative examples (respectively, images of different classes) away from each other, which is achieved due to Contrastive Loss:
C o n t r a s t i v e L o s s = log exp q · k + τ exp q · k + τ + i = 0 K 1 e x p q · k i τ ,
where q is the embedding for the query image, k + and k are the embeddings for positive and negative of the query image examples, respectively. τ is the temperature parameter; in this work, τ = 0.07 .
In this paper, the MoCo v2 strategy [50] is used as an algorithm for self-supervised pre-training. Figure 5 shows a scheme of this method. Its idea is the following:
  • Two augmentations are applied to the input image;
  • Then one augmented image goes to the encoder, while the other goes to the momentum encoder;
  • The outputs of the encoder and momentum encoder are embeddings that are used in the calculation of contrastive loss (two augmentations are taken as positive examples, while examples from the queue are considered as negative);
  • Embedding received from the momentum encoder is added to the end of the queue, and the queue is built according to the FIFO strategy;
  • The momentum encoder weights are updated by momentum averaging the encoder weights.
Thus, it seems possible to pre-train the neural network to distinguish people from each other without any markup. This task is somewhat simpler than person re-identification. However, it can be assumed that for the problem under consideration, it is more efficient to fine-tune a network that can somehow distinguish people, rather than classify various objects represented in ImageNet. This hypothesis is supported by the experimental evaluation presented in Section 4.3.4.

4. Experiments

4.1. Datasets

In the experiments conducted in this paper, three types of datasets were used: datasets designed for training and testing the algorithm for the person re-identification task, data for training and testing the Filter Module, as well as data for self-supervised pre-training. Next, all the datasets used in this work will be described.

4.1.1. Datasets for Person Re-ID

In this paper, the well-known datasets Market-1501 [1], DukeMTMC-reID [2], and MSMT17 [3] were used to train and test the proposed person Re-ID algorithms. All these datasets consist of full-body images of people taken from multiple cameras. For experiments related to upper-body re-identification, in this paper, we used modified versions of the datasets that were proposed in our previous work [57].

4.1.2. Datasets for Filter Module

Since the idea of filtering input data during the testing of the re-identification algorithm was not previously mentioned in other works, we had to face the problem associated with the lack of datasets for training and testing the Filter Module. For this reason, we decided to develop criteria for classifying images into correct and incorrect.
The image can be considered as correct if it meets the following conditions:
  • There should be a person in the image;
  • It should be obvious from the image which person it belongs to;
  • The desired person should not be overlapped too much by other people or objects;
  • The image should not be heavily cropped and the person in the image should not be too close;
  • The desired person and the colors of the clothes should be distinguishable in the image.
According to the above criteria, the dataset MSMT17 for person Re-ID was manually marked up. Figure 6 shows examples of incorrect images.

4.1.3. Datasets for Self-Supervised Pre-Training

As part of this work, 371 videos from surveillance cameras publicly broadcast on the Internet were selected, with a total duration of more than 2500 h. For 2.5 weeks, all videos were processed according to the fully automated data collection strategy described in Section 3.5.1, and as a result, about 11.5 million crops of about 980 thousand people were received (Figure 7).
Table 2 shows a comparative assessment of the obtained dataset for self-supervised pre-training with well-known datasets for the person Re-ID task.

4.2. Metrics

In the re-identification task, two metrics are most often used to determine the quality of algorithms: R a n k N and m A P . The first one is calculated as follows:
R a n k N = q Q A c c N ( q ) | Q | ,
where Q is the set of all query images involved in testing, and A c c N takes the following values:
A c c N ( q ) = 1 , if   the   Top - N   of   the   gallery   images   issued   by   the   algorithm includes   an   image   with   the   same   ID   as   the   image   q 0 , else
In the R a n k N metric, the parameter N means the number of gallery images with the highest confidence of the Re-ID algorithm for the query image, among which there should be an image with the desired person ID.
As for the mAP metric, it is calculated as follows:
m A P = q Q A P ( q ) | Q | ,
where A P ( q ) is the area under the Precision–Recall curve for the image q. Precision and Recall, in turn, are defined as:
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
where T P is True Positive matches, F P is False Positive matches, and F N is False Negative matches.

4.3. Experimental Results

There are two types of procedures for testing person re-identification algorithms: standard test protocol and cross-domain Re-ID. In the first case, we train and test the algorithm on the same dataset. In cross-domain Re-ID, we train the algorithm on one dataset and test it on another (the domain gap problem mentioned in Section 1).
Cross-domain Re-ID is the closest to the actual use of the algorithm in practice because, in practice, it is often impossible to train a neural network on the data on which it will be applied. Therefore, in this paper, during the experimental evaluation, the main emphasis was placed on tests in a scenario where the algorithm is trained on one dataset and tested on another.

4.3.1. Using Larger and Advanced Neural Networks as the Encoder

As can be seen from the experimental evaluation presented in Table 3, the hypothesis that larger and more advanced neural networks are better able to solve the re-identification problem, as put forward in Section 3.1, has been confirmed.
Table 4 presents an experimental evaluation showing the impact of using larger and more advanced neural network architectures on the quality of cross-domain Re-ID. As can be seen from the experimental results, the use of such neural networks is justified, especially in the case of cross-domain re-identification. Moreover, if we take into account the time measurements presented in Table 1, the use of Res2Net50 as the encoder in practice will not particularly affect the performance of the algorithm as a whole, but it will improve its quality in the Re-ID task.
Furthermore, from the results presented in Table 4, we can conclude that the use of larger datasets for training can significantly improve the results of person re-identification in the scenario when the algorithm is trained on one dataset and tested on another. This is very useful in practice, so in all the following experiments in this paper, training will be carried out with the large dataset MSMT17 Merged, and testing will be conducted with other smaller datasets.

4.3.2. Improving the Stability of the Algorithm to Image Changes and Noise

The experimental evaluation presented in Table 5 confirms the idea that the use of AugMix augmentation and JS loss reduces the sensitivity of the algorithm to image changes and noise and also improves the overall quality of solving the re-identification problem.
This result is important since most of the existing Re-ID algorithms are unstable with image changes, which often occur in practice due to detector errors, occlusions, and many other factors.

4.3.3. Filter Module

As mentioned earlier in Section 3.3, bad images can be fed to the input of the Re-ID algorithm, due to which the overall quality of re-identification may deteriorate. In this work, in the inference stage for preliminary filtering of such images, the Filter Module was proposed. In this series of experiments, the influence of the Filter Module, depending on the thresholds of the values of image suitability issued by it, on the quality of person Re-ID is considered.
As can be seen from the results of all experiments (Table 6 and Table 7), pre-filtering of the gallery and the set of query images has a positive effect on the metrics used in re-identification. Moreover, in most cases, we are talking about an increase in quality by several percent for both metrics, with a very small decrease in the volume of test sets. This suggests that the Filter Module does not just reject random images, but really looks for images unsuitable for person Re-ID and does not feed them to the input of the main Re-ID algorithm.
Thus, the Filter Module proposed in this paper can be used in practice to predict the suitability of images for re-identification and improve the quality of person Re-ID without a significant increase in computational costs due to the lightweight architecture of the Filter Module.

4.3.4. Upper-Body Re-ID and Self-Supervised Pre-Training

In this paper, in addition to the classical full-body re-identification, the algorithm based on upper-body detections is proposed (Section 3.4). The experimental results presented in Table 8 show that the use of Res2Net50 as the encoder, as well as the use of AugMix augmentation and JS Loss during algorithm training, can significantly improve the quality of upper-body Re-ID (as well as for full-body re-identification).
Furthermore, the conducted experiments show that self-supervised pre-training on the data collected using the proposed fully automated data collection strategy (Section 3.5 and Section 4.1.3) helps to improve the quality of upper-body Re-ID. Thus, the hypothesis put forward earlier in Section 3.5.2 was experimentally confirmed that if the neural network is pre-trained on data that are most similar to the target rather than images from ImageNet, then after fine-tuning this network on data for the final task, it becomes possible to achieve a better quality solution.
Moreover, during the experiments, two no less important and interesting observations were obtained for the problem of person re-identification:
  • Self-supervised pre-training on images of people allows us to speed up the neural network training procedure for re-identification by 1.5 times. That is, a result identical to the fine-tuning of the network pre-trained in supervised mode on ImageNet can be achieved in fewer epochs with the fine-tuning of the network pre-trained in self-supervised mode.
  • From the dataset described in Section 4.1.3, a subsample was randomly selected (about 300 thousand images of people). Further, the neural network was pre-trained on it in self-supervised mode. After that, it was found that fine-tuning the neural network pre-trained in self-supervised mode on a relatively small dataset can be more efficient in terms of re-identification quality than fine-tuning the network pre-trained in supervised mode on the whole ImageNet.
Thus, the upper-body Re-ID algorithm proposed in this paper, despite using a smaller area of the human body for analysis, has competitive results and is the preferred option in scenarios when it comes to video analytics on scenes with a large crowd of people.

5. Discussion

5.1. Comparison with Other Re-ID Methods

The main goal of our work is not to get ahead of the state-of-the-art re-identification algorithms but to study the main difficulties of the problem under consideration that arise in practice. As mentioned earlier, this paper proposes approaches (Section 3) that can be integrated with almost any of the popular modern person Re-ID algorithms in order to improve their quality while maintaining their computational complexity. On the example of a simple baseline [10], the consistency of all proposed approaches was demonstrated (Section 4.3).
However, despite the simplicity of the chosen baseline, with the help of the proposed approaches, it was possible to make it even more competitive in a cross-domain scenario with other more complex and advanced re-identification algorithms (Table 9). At the same time, the computational complexity of a simple baseline remained practically unchanged at the inference stage.
As can be judged from the experimental results presented in Table 9, the proposed approaches can significantly improve the quality of current algorithms while maintaining their computational complexity. Thus, using the methods proposed in this paper, it is possible to improve the quality of almost any of the popular modern algorithms for person re-identification. Demonstrating such an improvement is one of our future research goals (Section 5.4).

5.2. Practical Use of the Proposed Methods

Re-identification algorithms can be used in practice for their primary purpose—to search for people by query image in the database—and as components of other algorithms. For example, in [57,60,62], Re-ID algorithms are used as an integral part of the tracking algorithm for matching detections.
In our previous works [57,60], we also used re-identification methods to create a tracking algorithm capable of working with low detection frequency in order to solve the practically important task of estimating the waiting time in queues. That is, Re-ID methods can reduce the frequency of detection and thereby significantly reduce computational costs for the entire pipeline.
Returning to the primary purpose of re-identification algorithms, in most cases, the tracks of the movement of people are built on each surveillance camera. In most cases, this is completed in order to avoid storing a huge number of embeddings for each detection in the database, instead only a single embedding for each of the tracks, thereby facilitating the search. This approach leads to the question of how best to form this single embedding. The easiest way is to average the embeddings for all detections included in the track. However, poor detections (for example, due to detector errors or occlusions) can lead to poor performance of the Re-ID algorithm, and therefore averaging can work poorly. However, using the proposed Filter Module (Section 3.3) allows solving this dilemma in at least two ways:
  • When calculating the average embedding, it is possible to exclude those detections that have a low suitability value issued by the Filter Module;
  • Among all the detections in the track, choose the best shot. The best shot can be the detection that has the highest suitability value issued by the Filter Module. Furthermore, as a single embedding for the track, take the embedding corresponding to this detection.

5.3. Why Is Upper-Body Re-ID Useful?

If we compare the results presented in Table 5 and Table 8, it may seem that it is always worth using the full-body re-identification algorithm and that there is no sense in the upper-body Re-ID. However, in this paper, we propose the upper-body Re-ID algorithm as a result of our research aimed at practice.
In real scenes from surveillance cameras, occlusions are very frequent, because of which, in most cases, the full body is not visible, with the exception of only its upper part (Figure 4). In the case of occlusions, both detector errors are possible (for example, several people combine into one detection) and errors of the re-identification itself. Since well-known datasets for person Re-ID do not fully cover such cases (occlusions are much less common in them), it may seem that the full-body re-identification always works better.
Thus, it is recommended to combine these algorithms. If people are clearly visible in full height, use the full-body Re-ID algorithm; otherwise, the upper-body re-identification is recommended.

5.4. Future Research

There are many directions for continuing research on this topic. For example, it is necessary to integrate the proposed approaches into other existing Re-ID algorithms. Furthermore, the fully automated data collection strategy proposed in this paper implies obtaining some automatic markup through the use of tracking (Section 3.5.1). In this paper, this markup was not used during self-supervised pre-training. In theory, creating a self-supervised pre-training algorithm that will use this markup can improve the quality of the final algorithm. In conclusion, as noted in Section 5.3, well-known datasets for the problem under consideration do not reflect all the problems that may arise in practice, so there is a need to collect new datasets for a more honest comparison of the full-body re-identification and the upper-body Re-ID algorithms.

6. Conclusions

In this paper, the main problems that arise in the practical solution of the task of person re-identification were considered. We have proposed several approaches that allow us to improve the quality of existing Re-ID algorithms, as well as preserve its computational complexity. In summary, the contributions of this paper are concluded as follows:
  • We have shown that larger and advanced neural networks are able to solve the problem of person re-identification better. Furthermore, also in this paper, the encoder architecture was chosen, which improves the balance between the quality of the algorithm and its speed.
  • Approaches have been proposed that make it possible to increase the algorithm’s stability in response to image changes and noise.
  • We have proposed the new Filter Module designed to pre-filter input data before feeding the data to the main re-identification algorithm.
  • We have shown that due to occlusions, in practice, it is better to use upper-body detections instead of full-body detections for person Re-ID.
  • The fully automated data collection strategy from surveillance cameras for self-supervised pre-training has been proposed in order to increase the generality of neural networks on real-world data.
The experimental evaluation presented in this paper has shown the consistency of all the proposed approaches. Among other things, in this paper, the practical application of the proposed algorithms was considered. We believe that the results of this work can become an object for further research since they can be transferred to other computer vision tasks with a high degree of probability.

Author Contributions

Conceptualization, T.M. and D.K.; methodology, T.M. and D.K.; software, T.M.; validation, T.M. and D.K.; writing—original draft preparation, T.M, D.K. and A.K.; writing—review and editing, T.M., D.K. and A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviation is used in this manuscript:
Re-IDRe-identification

References

  1. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable person re-identification: A benchmark. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  2. Ristani, E.; Solera, F.; Zou, R.; Cucchiara, R.; Tomasi, C. Performance measures and a data set for multi-target, multi-camera tracking. In Proceedings of the Computer Vision–ECCV 2016 Workshops, Amsterdam, The Netherlands, 8–10, 15–16 October 2016; Part II. pp. 17–35. [Google Scholar]
  3. Wei, L.; Zhang, S.; Gao, W.; Tian, Q. Person transfer gan to bridge domain gap for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 79–88. [Google Scholar]
  4. Bai, Y.; Jiao, J.; Ce, W.; Liu, J.; Lou, Y.; Feng, X.; Duan, L.Y. Person30k: A dual-meta generalization network for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2123–2132. [Google Scholar]
  5. Song, C.; Huang, Y.; Ouyang, W.; Wang, L. Mask-guided contrastive attention model for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1179–1188. [Google Scholar]
  6. Cai, H.; Wang, Z.; Cheng, J. Multi-Scale Body-Part Mask Guided Attention for Person Re-identification. Available online: http://arxiv.org/abs/1904.11041 (accessed on 5 July 2023).
  7. Kalayeh, M.M.; Basaran, E.; Gökmen, M.; Kamasak, M.E.; Shah, M. Human semantic parsing for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1062–1071. [Google Scholar]
  8. Gao, H.; Chen, S.; Zhang, Z. Parts Semantic Segmentation Aware Representation Learning for Person Re-Identification. Appl. Sci. 2019, 9, 1239. [Google Scholar] [CrossRef]
  9. Wu, X.; Songlin, S.; Meixia, F. Person re-identification based on semantic segmentation. Signal Inf. Process. Netw. Computers. Lect. Notes Electr. Eng. 2020, 628, 903–909. [Google Scholar]
  10. Luo, H.; Gu, Y.; Liao, X.; Lai, S.; Jiang, W. Bag of tricks and a strong baseline for deep person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1487–1495. [Google Scholar]
  11. Wang, G.; Yuan, Y.; Chen, X.; Li, J.; Zhou, X. Learning discriminative features with multiple granularities for person re-identification. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 274–282. [Google Scholar]
  12. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 3001–13008. [Google Scholar]
  13. Yan, C.; Pang, G.; Bai, X.; Liu, C.; Ning, X.; Gu, L.; Zhou, J. Beyond triplet loss: Person re-identification with fine-grained difference-aware pairwise loss. IEEE Trans. Multimed. 2021, 24, 1665–1677. [Google Scholar] [CrossRef]
  14. Gu, H. Autoloss-gms: Searching generalized margin-based softmax loss function for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–23 June 2022; pp. 4744–4753. [Google Scholar]
  15. Zhou, X. Adaptive Sparse Pairwise Loss for Object Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 19691–19701. [Google Scholar]
  16. Zhang, T.; Xie, L.; Wei, L.; Zhuang, Z.; Zhang, Y.; Li, B.; Tian, Q. Unrealperson: An adaptive pipeline towards costless person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11506–11515. [Google Scholar]
  17. Xiang, S.; Fu, Y.; You, G.; Liu, T. Taking a closer look at synthesis: Fine-grained attribute analysis for person re-identification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2021), Toronto, ON, Canada, 6–11 June 2021; pp. 3765–3769. [Google Scholar]
  18. Hendrycks, D.; Mu, N.; Cubuk, E.D.; Zoph, B.; Gilmer, J.; Lakshminarayanan, B. Augmix: A Simple Data Processing Method to Improve Robustness and Uncertainty. Available online: https://arxiv.org/abs/1912.02781 (accessed on 5 July 2023).
  19. Zhou, K. Omni-scale feature learning for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3702–3712. [Google Scholar]
  20. Chen, X.; Fu, C.; Zhao, Y.; Zheng, F.; Song, J.; Ji, R.; Yang, Y. Salience-Guided Cascaded Suppression Network for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3300–3310. [Google Scholar]
  21. Liao, S.; Shao, L. Interpretable and generalizable person re-identification with query-adaptive convolution and temporal lifting. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 456–474. [Google Scholar]
  22. Liao, S.; Shao, L. Graph sampling based deep metric learning for generalizable person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–23 June 2022; pp. 7359–7368. [Google Scholar]
  23. He, S.; Luo, H.; Wang, P.; Wang, F.; Li, H.; Jiang, W. Transreid: Transformer-based object re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 15013–15022. [Google Scholar]
  24. Liao, S.; Shao, L. Transmatcher: Deep image matching through transformers for generalizable person re-identification. Adv. Neural Inf. Process. Syst. 2021, 34, 1992–2003. [Google Scholar]
  25. Tan, L.; Dai, P.; Ji, R.; Wu, Y. Dynamic prototype mask for occluded person re-identification. In Proceedings of the 30th ACM International Conference on Multimedia, Ottawa, ON, Canada, 10–14 October 2022; pp. 531–540. [Google Scholar]
  26. Zhang, G.; Zhang, Y.; Zhang, T.; Li, B.; Pu, S. PHA: Patch-Wise High-Frequency Augmentation for Transformer-Based Person Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 14133–14142. [Google Scholar]
  27. Chen, Y.; Zhu, X.; Gong, S. Instance-Guided Context Rendering for Cross-Domain Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 232–242. [Google Scholar]
  28. Luo, C.; Song, C.; Zhang, Z. Generalizing Person Re-Identification by Camera-Aware Invariance Learning and Cross-Domain Mixup. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Swizterland, 2020; Volume 12360, pp. 224–241. [Google Scholar]
  29. Ge, Y.; Zhu, F.; Chen, D.; Zhao, R.; Li, H. Self-Paced Contrastive Learning with Hybrid Memory for Domain Adaptive Object Re-ID. In Proceedings of the Neural Information Processing Systems, Virtual, 6–12 December 2020; Volume 33, pp. 11309–11321. [Google Scholar]
  30. Chen, H.; Lagadec, B.; Bremond, F. Ice: Inter-instance contrastive encoding for unsupervised person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 14960–14969. [Google Scholar]
  31. Dai, Z.; Wang, G.; Yuan, W.; Zhu, S.; Tan, P. Cluster contrast for unsupervised person re-identification. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 1142–1160. [Google Scholar]
  32. Zhang, X.; Li, D.; Wang, Z.; Wang, J.; Ding, E.; Shi, J.Q.; Zhang, Z.; Wang, J. Implicit sample extension for unsupervised person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–23 June 2022; pp. 7369–7378. [Google Scholar]
  33. Qian, X.; Wang, W.; Zhang, L.; Zhu, F.; Fu, Y.; Xiang, T.; Jiang, Y.G.; Xue, X. Long-term cloth-changing person re-identification. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–3 December 2020; pp. 71–88. [Google Scholar]
  34. Yang, Q.; Wu, A.; Zheng, W.S. Person re-identification by contour sketch under moderate clothing change. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 2029–2046. [Google Scholar] [CrossRef] [PubMed]
  35. Hong, P.; Wu, T.; Wu, A.; Han, X.; Zheng, W.S. Fine-grained shape-appearance mutual learning for cloth-changing person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10513–10522. [Google Scholar]
  36. Jin, X.; He, T.; Zheng, K.; Yin, Z.; Shen, X.; Huang, Z.; Feng, R.; Huang, J.; Chen, Z.; Hua, X.S. Cloth-changing person re-identification from a single image with gait prediction and regularization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–23 June 2022; pp. 14278–14287. [Google Scholar]
  37. Chao, H.; He, Y.; Zhang, J.; Feng, J. Gaitset: Regarding gait as a set for cross-view gait recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 8126–8133. [Google Scholar]
  38. Zhang, Z.; Tran, L.; Yin, X.; Atoum, Y.; Liu, X.; Wan, J.; Wang, N. Gait recognition via disentangled representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4710–4719. [Google Scholar]
  39. Jiang, J.; Jin, K.; Qi, M.; Wang, Q.; Wu, J.; Chen, C. A cross-modal multi-granularity attention network for RGB-IR person re-identification. Neurocomputing 2020, 406, 59–67. [Google Scholar] [CrossRef]
  40. Liu, H.; Cheng, J.; Wang, W.; Su, Y.; Bai, H. Enhancing the discriminative feature learning for visible-thermal cross-modality person re-identification. Neurocomputing 2020, 398, 11–19. [Google Scholar] [CrossRef]
  41. Wu, A.; Zheng, W.S.; Gong, S.; Lai, J. RGB-IR person re-identification by cross-modality similarity preservation. Int. J. Comput. Vis. 2020, 128, 1765–1785. [Google Scholar] [CrossRef]
  42. Josi, A.; Alehdaghi, M.; Cruz, R.M.; Granger, E. Multimodal data augmentation for visual-infrared person ReID with corrupted data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 32–41. [Google Scholar]
  43. Kim, M.; Kim, S.; Park, J.; Park, S.; Sohn, K. PartMix: Regularization Strategy to Learn Part Discovery for Visible-Infrared Person Re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 18621–18632. [Google Scholar]
  44. Bachman, P.; Hjelm, R.D.; Buchwalter, W. Learning representations by maximizing mutual information across views. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 15535–15545. [Google Scholar]
  45. Misra, I.; Maaten, L.V.D. Self-supervised learning of pretext-invariant representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6707–6717. [Google Scholar]
  46. Zbontar, J.; Jing, L.; Misra, I.; LeCun, Y.; Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 12310–12320. [Google Scholar]
  47. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 12–18 July 2020; pp. 1597–1607. [Google Scholar]
  48. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 3733–3742. [Google Scholar]
  49. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  50. Chen, X.; Fan, H.; Girshick, R.; He, K. Improved Baselines with Momentum Contrastive Learning. Available online: https://arxiv.org/abs/2003.04297 (accessed on 5 July 2023).
  51. Bardes, A.; Ponce, J.; LeCun, Y. Vicreg: Variance-Invariance-Covariance Regularization for Self-Supervised Learning. Available online: https://arxiv.org/abs/2105.04906 (accessed on 11 August 2023).
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  53. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef] [PubMed]
  54. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  55. ONNX: Open Neural Network Exchange. Available online: https://github.com/onnx/onnx (accessed on 11 August 2023).
  56. ONNX Runtime. Available online: https://onnxruntime.ai (accessed on 11 August 2023).
  57. Mamedov, T.; Kuplyakov, D.; Konushin, A. Queue Waiting Time Estimation Using Person Re-identification by Upper Body. In Proceedings of the 31th International Conference on Computer Graphics and Machine Vision, Nizhny Novgorod, Russia, 27–30 September 2021; pp. 464–474. [Google Scholar]
  58. Dong, X.; Shen, J. Triplet loss in siamese network for object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 459–474. [Google Scholar]
  59. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  60. Mamedov, T.Z.; Kuplyakov, D.A.; Konushin, A.S. Video Analytics Using Detection on Sparse Frames. Program. Comput. Softw. 2022, 48, 155–163. [Google Scholar] [CrossRef]
  61. Zhou, K.; Yang, Y.; Cavallaro, A.; Xiang, T. Learning generalisable omni-scale representations for person re-identification. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 444, 5056–5069. [Google Scholar] [CrossRef] [PubMed]
  62. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
Figure 1. Scheme of the baseline. At the training stage, the person classification problem is solved. At the inference stage, only embeddings obtained after the batch normalization layer are used to compare images using cosine similarity.
Figure 1. Scheme of the baseline. At the training stage, the person classification problem is solved. At the inference stage, only embeddings obtained after the batch normalization layer are used to compare images using cosine similarity.
Sensors 23 07382 g001
Figure 2. An example of a sequence of images from the track of a girl in a black sweatshirt. Incorrect images are highlighted in red, while correct images are not highlighted. Explanations: the eighth image is marked as incorrect since it is difficult to identify the person you are looking for. The twelfth image is highlighted as incorrect since the desired person is strongly occluded by another person. There is a high probability of an error in the re-identification algorithm on these images, so it is worth evaluating the suitability of each of the images for subsequent analysis in advance.
Figure 2. An example of a sequence of images from the track of a girl in a black sweatshirt. Incorrect images are highlighted in red, while correct images are not highlighted. Explanations: the eighth image is marked as incorrect since it is difficult to identify the person you are looking for. The twelfth image is highlighted as incorrect since the desired person is strongly occluded by another person. There is a high probability of an error in the re-identification algorithm on these images, so it is worth evaluating the suitability of each of the images for subsequent analysis in advance.
Sensors 23 07382 g002
Figure 3. Scheme of the proposed Filter Module. This neural network has a simple architecture since it performs a secondary function and should not significantly increase computational costs of the entire pipeline in practice.
Figure 3. Scheme of the proposed Filter Module. This neural network has a simple architecture since it performs a secondary function and should not significantly increase computational costs of the entire pipeline in practice.
Sensors 23 07382 g003
Figure 4. An example of a frame from a surveillance camera demonstrating the fact that upper-body detections have greater visibility than full-body detections (bounding boxes are drawn for clarity).
Figure 4. An example of a frame from a surveillance camera demonstrating the fact that upper-body detections have greater visibility than full-body detections (bounding boxes are drawn for clarity).
Sensors 23 07382 g004
Figure 5. Scheme of the MoCo v2 strategy.
Figure 5. Scheme of the MoCo v2 strategy.
Sensors 23 07382 g005
Figure 6. Examples of incorrect images on which there is a high probability of an error of the re-identification algorithm. In the first image, there is no person; in the second, there are two people at once (i.e., it is not clear which of them needs to be re-identified); in the third, only a part of the body, which is very difficult to re-identify, got into the image; in the fourth, the image is too dark, so it is almost impossible to distinguish the person and the colours of clothes.
Figure 6. Examples of incorrect images on which there is a high probability of an error of the re-identification algorithm. In the first image, there is no person; in the second, there are two people at once (i.e., it is not clear which of them needs to be re-identified); in the third, only a part of the body, which is very difficult to re-identify, got into the image; in the fourth, the image is too dark, so it is almost impossible to distinguish the person and the colours of clothes.
Sensors 23 07382 g006
Figure 7. Examples of images in the collected dataset for self-supervised pre-training.
Figure 7. Examples of images in the collected dataset for self-supervised pre-training.
Sensors 23 07382 g007
Table 1. The time measurements for ResNet50 and Res2Net50 on the Intel Core i5-10600K.
Table 1. The time measurements for ResNet50 and Res2Net50 on the Intel Core i5-10600K.
EncoderSingle Core Speed *
ResNet50202 ms
Res2Net50224 ms
* To calculate the speed of the neural network on single core, we convert our model to ONNX [55] and run it 1000 times on one processor core using ONNX Runtime [56]. After that, we find the average speed of the neural network.
Table 2. Comparison of the collected dataset for self-supervised pre-training with datasets for person re-identifcation.
Table 2. Comparison of the collected dataset for self-supervised pre-training with datasets for person re-identifcation.
DatasetImagesPersonsLocationSeasonsCameraResolution
Market-150132,6681501UniversityFixedFixedFixed
DukeMTMC-reID36,4111812UniversityFixedFixedFixed
MSMT17126,4414101UniversityFixedFixedFixed
Our>11.5 M>980 KDifferentDifferentDifferentDifferent
Table 3. Experimental evaluation of the algorithm with different backbones (standard test protocol).
Table 3. Experimental evaluation of the algorithm with different backbones (standard test protocol).
AlgorithmMarket-1501DukeMTMC-reID
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline (ResNet50)94.5085.9086.4076.40
With Res2Net5095.0187.1988.1577.61
The best results by metrics are highlighted in bold.
Table 4. Experimental evaluation of the algorithm with different backbones (cross-domain Re-ID). Symbols: D—DukeMTMC-reID; M—Market-1501; MSMT17 Merged—MSMT17, which combines training and test parts.
Table 4. Experimental evaluation of the algorithm with different backbones (cross-domain Re-ID). Symbols: D—DukeMTMC-reID; M—Market-1501; MSMT17 Merged—MSMT17, which combines training and test parts.
AlgorithmD → MMSMT17 → M
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline (ResNet50)54.3025.5058.8230.25
With Res2Net5054.1625.9562.0232.70
AlgorithmM → DMSMT17 → D
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline (ResNet50)41.4025.7058.5338.25
With Res2Net5044.8426.6763.6042.59
AlgorithmMSMT17 Merged → MMSMT17 Merged → D
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline (ResNet50)65.6537.7166.1647.74
With Res2Net5069.7441.3270.6552.37
The best results by metrics are highlighted in bold.
Table 5. Experimental evaluation of the algorithm with AugMix augmentation and JS Loss (cross-domain Re-ID). Symbols: M—Market-1501; D—DukeMTMC-reID; MSMT17 Merged—MSMT17, which combines training and test parts.
Table 5. Experimental evaluation of the algorithm with AugMix augmentation and JS Loss (cross-domain Re-ID). Symbols: M—Market-1501; D—DukeMTMC-reID; MSMT17 Merged—MSMT17, which combines training and test parts.
AlgorithmMSMT17 Merged → MMSMT17 Merged → D
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline65.6537.7166.1647.74
+Res2Net5069.7441.3270.6552.37
+AugMix and JS Loss71.0843.3972.5153.71
The best results by metrics are highlighted in bold.
Table 6. Experimental evaluation of the dependence of the quality of re-identification on the probability threshold applied to the Filter Module: the case when the neural network for Re-ID is trained on MSMT17 Merged and tested on Market-1501 (cross-domain Re-ID).
Table 6. Experimental evaluation of the dependence of the quality of re-identification on the probability threshold applied to the Filter Module: the case when the neural network for Re-ID is trained on MSMT17 Merged and tested on Market-1501 (cross-domain Re-ID).
Threshold Rank 1 ( ) , % mAP ( ) , %Gallery Reduction, %Query Reduction, %
0.071.0843.390.000.00
0.171.0843.390.010.00
0.271.1343.430.250.06
0.371.1543.460.800.09
0.471.1943.511.470.15
0.571.2343.622.290.50
0.671.3943.843.501.34
0.771.5544.135.402.38
0.872.1744.878.204.81
0.973.6946.8114.0411.28
Table 7. Experimental evaluation of the dependence of the quality of re-identification on the probability threshold applied to the Filter Module: the case when the neural network for Re-ID is trained on MSMT17 Merged and tested on DukeMTMC-reID (cross-domain Re-ID).
Table 7. Experimental evaluation of the dependence of the quality of re-identification on the probability threshold applied to the Filter Module: the case when the neural network for Re-ID is trained on MSMT17 Merged and tested on DukeMTMC-reID (cross-domain Re-ID).
Threshold Rank 1 ( ) , % mAP ( ) , %Gallery Reduction, %Query Reduction, %
0.072.5153.710.000.00
0.172.5453.730.010.04
0.272.7754.060.250.49
0.373.1354.640.801.12
0.473.4055.111.471.75
0.573.8255.642.292.51
0.674.1256.203.503.23
0.774.8057.075.404.44
0.875.6858.408.206.91
0.977.5560.9914.0412.30
Table 8. Experiments with upper-body Re-ID and self-supervised pre-training (cross-domain Re-ID). Symbols: MSMT17 Merged—MSMT17, which combines training and test parts.
Table 8. Experiments with upper-body Re-ID and self-supervised pre-training (cross-domain Re-ID). Symbols: MSMT17 Merged—MSMT17, which combines training and test parts.
AlgorithmMSMT17 Merged → DukeMTMC-reID
Rank 1 ( ) , % mAP ( ) , %
Baseline (upper-body)50.9733.57
+Res2Net5053.3434.73
+AugMix and JS Loss57.1237.69
+Self-supervised Pre-training58.2438.70
The best results by metrics are highlighted in bold.
Table 9. Experimental comparison of full-body re-identification algorithms (cross-domain Re-ID). Symbols: M—Market-1501; D—DukeMTMC-reID; MSMT17 Merged—MSMT17, which combines training and test parts.
Table 9. Experimental comparison of full-body re-identification algorithms (cross-domain Re-ID). Symbols: M—Market-1501; D—DukeMTMC-reID; MSMT17 Merged—MSMT17, which combines training and test parts.
AlgorithmMSMT17 Merged → MMSMT17 Merged → D
Rank 1 ( ) , % mAP ( ) , % Rank 1 ( ) , % mAP ( ) , %
Baseline [10]65.6537.7166.1647.74
OSNet-IBN [19]66.4737.2267.4345.59
OSNet-AIN [61]70.1343.2871.0652.72
QAConv [21]72.6443.11
Our71.0843.3972.5153.71
Our + Filter Module73.6946.8177.5560.99
The best results by metrics are highlighted in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mamedov, T.; Kuplyakov, D.; Konushin, A. Approaches to Improve the Quality of Person Re-Identification for Practical Use. Sensors 2023, 23, 7382. https://doi.org/10.3390/s23177382

AMA Style

Mamedov T, Kuplyakov D, Konushin A. Approaches to Improve the Quality of Person Re-Identification for Practical Use. Sensors. 2023; 23(17):7382. https://doi.org/10.3390/s23177382

Chicago/Turabian Style

Mamedov, Timur, Denis Kuplyakov, and Anton Konushin. 2023. "Approaches to Improve the Quality of Person Re-Identification for Practical Use" Sensors 23, no. 17: 7382. https://doi.org/10.3390/s23177382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop