Next Article in Journal
Regime Map of the Effective Medium Approximation Modelling of Micro-Rough Surfaces in Ellipsometry
Previous Article in Journal
Design of a Compact and Minimalistic Intermediate Phase Shifting Feed Network for Ka-Band Electrical Beam Steering
Previous Article in Special Issue
Approaches to Improve the Quality of Person Re-Identification for Practical Use
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generative Approach to Person Reidentification

Department of Informatics—Science and Engineering (DISI), University of Bologna, 40126 Bologna, Italy
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1240; https://doi.org/10.3390/s24041240
Submission received: 21 December 2023 / Revised: 25 January 2024 / Accepted: 8 February 2024 / Published: 15 February 2024
(This article belongs to the Special Issue Person Re-Identification Based on Computer Vision)

Abstract

:
Person Re-identification is the task of recognizing comparable subjects across a network of nonoverlapping cameras. This is typically achieved by extracting from the source image a vector of characteristic features of the specific person captured by the camera. Learning a good set of robust, invariant and discriminative features is a complex task, often leveraging contrastive learning. In this article, we explore a different approach, learning the representation of an individual as the conditioning information required to generate images of the specific person starting from random noise. In this way we decouple the identity of the individual from any other information relative to a specific instance (pose, background, etc.), allowing interesting transformations from one identity to another. As generative models, we use the recent diffusion models that have already proven their sensibility to conditioning in many different contexts. The results presented in this article serve as a proof-of-concept. While our current performance on common benchmarks is lower than state-of-the-art techniques, the approach is intriguing and rich of innovative insights, suggesting a wide range of potential improvements along various lines of investigation.

1. Introduction

The challenge of person re-identification lies in recognizing comparable subjects across a network of non-overlapping cameras, which is a common scenario in multi-camera surveillance systems [1]. In its typical formulation, the person re-ID task aims to identify a specific individual from an extensive collection of person images, known as the gallery, by utilizing a query image [2]. While this task falls under the framework of image retrieval problems, its unique objective of determining the identity of the person within a query image, typically expressed by a distinctive ID label, introduces intriguing peculiarities.
At its core, person re-ID involves acquiring knowledge about distinct characteristics of individuals, enabling the differentiation between images depicting the same person and those featuring different individuals. The difficulty of this task stems from substantial variations in viewpoint, pose, lighting, and image quality across diverse cameras in real-world scenarios. In such situations, individuals may appear in multiple cameras across various locations, thereby intensifying the challenge of feature acquisition [3].
The typical approach to the re-ID task involves a two-phase process: a representation learning phase, where interesting and distinctive features of individuals are extracted, and a metric learning phase, where training objectives are designed using various loss functions or sampling strategies. This process aims to shape the embedding space in a way suitable for searching and retrieval.
Both phases are complex and very interesting. For feature extraction, modern approaches have leveraged deep learning techniques, exploiting convolutional networks [4,5], Generative Adversarial Networks [6,7,8,9], Visual Transformers [10,11,12,13,14], and different kinds of attention mechanisms [15,16,17,18]. Features can be enhanced by exploiting part-based methods, focusing on extracting features from specific regions of interest [10,19,20], pretraining [21,22], and multitasking [23] via a suitable set of pretext tasks.
Metric Learning is typically based on some form of contrastive learning [24], either using contrastive loss [25,26,27,28] or, more frequently, triplet loss [29,30,31,32]. Both approaches aim to encourage similar examples to be close in the feature space and dissimilar examples to be separated. In the context of triplet loss, ref. [33] introduced a focal loss to the triplet criterion, assigning more weight to negative samples compared to positives. Additionally, computing centroids, such as the mean of all samples belonging to each instance, helps with cluster creations as well [34,35,36,37]. The triplets can be extended to a quadruplet loss, with two negative samples and one positive pair, to enlarge the inter-class variations [38] or to N-tuple loss for joint optimization of multiple instances [39]. An extensive amount of works use a form of contrastive loss in combination with the softmax loss [40,41], i.e., recent works are [19,23,42]. Contrastive learning can also be used in combination with techniques such as generative learning [43], multi-instance learning [44], and spatial attention [45].
Contrary to all such methods, we adopted a distinct approach by associating a unique latent representation with each identity. This representation was learned as the conditioning information necessary to generate diversified samples of the particular person from random Gaussian noise (see Figure 1a). The generator takes as input a (learned) embedding of the identity e m b ( i d ) , a noise vector z, and produces a sample image G E N ( e m b ( i d ) , z ) . The next step involves computing a right inverse G E N 1 to the generator, producing, from an individual’s image X, an embedded identity i d X and a noise vector z X , such that G E N ( i d X , z X ) = X (see Figure 1b). The idea is that i d X should capture the interesting and individual features of the person represented in X, while the “noise” z X should contain all information unrelated to the specific identity, such as background, pose, etc. The identity i d X is then used to retrieve samples from the gallery by comparing their embeddings in the latent space.
The inversion of the generator was the most delicate part of the work and is discussed in Section 4.
As the generative model, we employed an instance of the recent diffusion models [46,47] due to their remarkable sensitivity to conditioning, unbiased sampling capabilities, and excellent expressiveness. Examples of the generation of images conditioned on the individual identity are given in Figure 2.
The generative quality is not perfect, and can probably be improved upon. In particular, for this prototyping work, we used a low-resolution version of the dataset, with images of a 64 × 32 dimension (instead of the usual 128 × 64 ). The embedding space for the identities has 32 dimensions. In spite of being so small, the conditioning seems to work well, as exemplified in Figure 2.
The results presented in this article represent a preliminary investigation in the direction described above. They are primarily intended as a proof of concept, testing the feasibility of the approach. The current performance, measured on standard benchmarks such as Market-1501, is encouraging, though it falls short of various state-of-the-art approaches.

1.1. Achievements

The goal of this work was to answer two conceptual questions, namely:
  • Is it possible to characterize a person’s identity (from the person re-ID perspective) as the conditional information required to generate a diversified set of images relative to the given person?
  • In that case, is it possible to compute a right inverse to the generator, extracting from an image of a person their conditioning identity?
Although, from the point of view of person re-identification, our results are slightly below the state of the art, our work provides evidence that both questions are likely to have a positive answer.

1.2. Structure of the Work

The article is structured as follows. Section 2 discusses related work. Section 3 introduces the specific class of generative models utilized in this study, namely Denoising Diffusion Models, with a focus on a particular subclass known as Implicit Models. The methodology, including the inversion of the generative network, is detailed in Section 4. Section 5 provides a comprehensive description of the neural network architecture. A brief summary of our results is presented in Section 6. Section 7 explores the latent space of identities, examining the spatial organization of latent representations and the significance of different variation factors. Section 8 is dedicated to ablation studies and discusses alternative solutions that were considered but ultimately not adopted. Concluding remarks and directions for future research are provided in Section 9.

2. Related Work

In the introduction, we provided an overview of the primary methodologies used in addressing the person re-identification task. Given that this paper adopts a generative paradigm, this section focuses on exploring relevant literature within this specific context.
Generative techniques in person re-identification have primarily been utilized for data augmentation and unsupervised domain adaptation (UDA). The majority of these works employ Generative Adversarial Networks (GANs), which have only recently been challenged by the emergence of diffusion models.
Domain adaptation methods utilize style transformation techniques to adapt images from a source domain to resemble those of a target domain. This transformation allows the images to retain the content of the source domain while adopting the visual style of the target domain. The new images obtained in this way are then used to fine-tune the network parameters. GAN-based image style transfer has quickly become a popular method for transferring knowledge between source and target domains in unsupervised cross-domain person re-identification, as evidenced by [6,7,48,49].
CycleGANs, as discussed in [9,50], are employed for this purpose. In the first mentioned work, a target encoder is used to obtain a discriminative mapping of target images into the transformed source feature space, thereby fooling a domain discriminator. The role of the discriminator is to distinguish between the features of the target domain and those of the transformed source domain. In [50], the objective is to swap features of the source and target environments to generate cross-domain images that preserve identity-related features associated with the source (or target) background features. This process is subsequently reversed to reconstruct the original input image, facilitating a self-supervised cyclic generation.
Chen et al. [43] propose an alternative approach in their study. They employ a 3D mesh-guided GAN to generate views, which are then integrated with original images through memory-based contrastive learning, aiming to achieve view-invariant representations of instances. In this methodology, training occurs jointly; the GAN learns the data distribution through adversarial training, while the contrastive instance discriminator acquires representations by retrieving each instance from a pool of candidates.
In [8], GANs fulfill a dual role. They utilize CycleGAN and Siamese networks to transfer image styles between domains while ensuring pedestrian identities are preserved. This is followed by iterative self-training with GANs to enhance both global and local features within the target domain, thereby facilitating robust feature learning in unlabeled data.
In [51], the introduction of the PCDS-GAN model marks a significant advancement in synthesizing source-labeled images against domain-specific backgrounds, effectively bridging the domain gap and enhancing domain adaptation efficiency. This is accomplished by separating pedestrian images into foreground, background, and style features. A U-Net-based Hole-Filling Module (HFM) is employed for transferring the background between different domains, tasked with reconstructing the scene areas previously occupied by the source foreground. This technique enables the generation of person images set against a variety of target domain backgrounds.

3. Denoising Diffusion Models

In this section, we provide a concise overview of diffusion models, a topic extensively covered in existing literature. Readers familiar with this area may opt to skip this section.
Denoising Diffusion Models (DDMs) [46] mark a significant breakthrough in deep generative modeling, posing a challenge to the longstanding predominance of Generative Adversarial Networks [52]. These models have been applied with great success in various recent and notable projects, such as [53,54,55]. Key factors contributing to their rising prominence include unparalleled generation quality, high adaptability and ease of conditioning, the ability to generate diverse and robust samples, stable training processes, and remarkable scalability.
An important characteristic of this generative paradigm is its robust probabilistic foundation. However, a formal exploration of the underlying theory behind denoising diffusion models is beyond the scope of this article. For a deeper theoretical understanding, we refer the reader to the extensive literature available [46,47]. Here, we aim to provide an operationally focused description of the denoising model, comprehensible even without a deep understanding of its theoretical background.
In very rough terms, a diffusion model trains a single network to denoise images with a parametric amount of noise. This network is then used to generate new samples in an iterative manner, starting from a purely ’noisy’ image and progressively removing a decreasing amount of noise.
This process is traditionally referred to as reverse diffusion, as it aims to “invert” the direct diffusion process, which consists of iteratively adding noise (see Figure 3).
We concentrated on a particular subclass of diffusion models, known as Implicit Diffusion Models [47], which have been a focal point of our research in numerous previous works. A key feature of these models is their fully deterministic reverse diffusion process. This attribute is crucial for applications that require embedding the output back into its latent representation, as discussed in [57]. Another significant advantage of Implicit Diffusion Models is their efficiency, typically requiring only a minimal number of iterations (about 10, as reported in [58,59])—a stark contrast to other techniques that may need thousands of iterations. Additionally, their remarkable conditioning capabilities have been demonstrated in the field of precipitation forecasting [56], where the goal is to predict the probability distribution of a specific atmospheric parameter based on recent weather conditions.

3.1. The Denoising Network

The only trainable component of the reverse diffusion process is a denoising network  ϵ θ ( x t , α t , c ) , which takes as input a noisy image x t , a signal rate α t , possibly a condition c, and tries to guess the noise present in the image. The model is trained in a completely supervised way. The main steps are the following:
  • pick a random image x 0 from the train set, coherent with the condition c;
  • select a random step t of the process; to each step t is associated a signal rate α t defined by a suitable noise scheduling (more in the sampling section);
  • sample a random Gaussian noise ϵ ;
  • create a corrupted image as a weighted combination of x 0 and ϵ :
    x t = α t x 0 + 1 α t ϵ
  • train the network to properly guess the amount of noise present in x t , by minimizing the distance between ϵ ( x t , α t , c ) and ϵ .
The previous steps are summarized in the pseudocode of Algorithm 1, where q ( x 0 | c ) denotes the distribution of training data, relative to the condition c.
Algorithm 1 Training
1: repeat
2:        x 0 q ( x 0 | c )                 ▹ take a sample coherent with c
3:        t Uniform(1,…,T)                   ▹ choose a timestep
4:        ϵ N ( 0 ; I )                 ▹ create random Gaussian noise
5:        x t = α t x 0 + 1 α t ϵ         ▹ corrupt the sample with signal rate α t
6:       Take a gradient descent step on | | ϵ ϵ θ ( x t , α t , c ) | | 2   ▹ backpropagate the loss
7: until converged
The approach to conditional training previously described was formally examined in [60]. The paper proposes a blend of conditional and unconditional generation, a method some researchers found beneficial for training, though we observed no significant improvements.
Another strategy for implementing conditioning involves using an auxiliary classifier [52], akin in concept to AC-GANs [61]. This method trains a classifier f ϕ ( c | x t ) on a noisy image x t to predict its class c. The gradient x log f ϕ ( c | x t ) is then utilized to steer the diffusion sampling process, enhancing the generation of samples that more accurately reflect the given condition. This technique remains untested in our work but represents a potential avenue for future enhancements.
The preferred architecture for the denoising network adopts a U-Net structure [62]. Our specific version is discussed in Section 5.
To heighten the network’s sensitivity to the signal rate α t , this parameter is often encoded through a bespoke sinusoidal transformation, dividing it into a series of frequencies, mirroring the technique of positional encodings in Transformers [63].

3.2. Sampling

Sampling is an iterative process. Starting from a purely noisy image x T , we progressively remove noise by calling the denoising network. Specifically, if the error predicted by the network at timestep t is ϵ = ϵ ( x t , α t , c ) , then the corresponding denoised prediction is
x 0 ^ = 1 α t ( x t 1 α t ϵ )
Next, ϵ is re-injected into the network at signal rate α t 1 , to obtain the next noisy image:
x t 1 = α t 1 x ˜ 0 + 1 α t 1 ϵ
The pseudocode of the sampling process is given in Algorithm 2.
Algorithm 2 Sampling
1: x T N ( 0 , I )
2: for  t = T , . . . , 1  do
3:       ϵ = ϵ θ ( x t , α t , c )                 ▹ predict noise
4:       x ˜ 0 = 1 α t ( x t 1 α t ϵ )      ▹ compute denoised result
5:       x t 1 = α t 1 x ˜ 0 + 1 α t 1 ϵ   ▹ re-inject noise at rate α t 1
6: end for
A significant component of the reverse diffusion process is the scheduling of diffusion noise α t t = 1 T . The authors in [46] suggested the use of linear or quadratic schedules. However, such choices lead to a rapid decrease in the initial steps, creating challenges in the generation phase. The literature offers alternative scheduling functions that decrease more gradually, including the “cosine” or “continuous cosine” schedule [64,65]. Opting for a gentler scheduler not only solves the generation issues but also reduces the number of iterations needed. For our purposes, we have implemented a generator with 10 diffusion steps.

4. Methodology

As introduced earlier, the core concept is to assign a unique latent representation to each identity, which is learned as the necessary conditioning information for generating specific individual samples from random Gaussian noise. The overall structure of the generative model is depicted in Figure 1a. This model is an example of a diffusion model, as discussed in the preceding section.
The subsequent step involves reversing the generative model, as shown in Figure 1b. This process enables us to derive, from an individual’s image, the latent representation of their identity and noise that captures all details unrelated to identity, such as background and pose. The latent identity holds significance for person re-identification because comparing it with the latent representations of gallery samples allows us to identify the most similar identities.
The feasibility of inverting diffusion models through appropriately trained neural networks has been explored in many of our prior works. The overall idea was presented in [57]; this approach was applied in [58] for the “reification” of artistic portraits by embedding a portrait into a latent space of human faces and reconstructing the closest real approximation. In [59], the diffusion inversion was utilized to create a trajectory in the latent space that induces a smooth rotation effect on human faces.
In this work, after several trials, we decided to divide the inversion networks into two separate models, as illustrated in Figure 4. One model, named Img2ID, converts images into latent identities, while the other, named Img2Noise, generates the noise. Given that both models extract orthogonal features from the source image, there was no compelling justification for them to share layers.
There are a few different possibilities for training the inversion models. Suppose we have a labelled image i d , X i d . We can train Img2ID to minimize the distance:
e m b ( i d ) , i m g 2 I D ( X i d )
where e m b is the identity embedder of the generator. Similarly, Img2Noise can be trained to minimize the error:
G E N ( I m g 2 N o i s e ( X i d ) , e m b ( i d ) ) , X i d
as the noise synthesized by I m g 2 N o i s e ( X i d ) should result in the generation of X i d when conditioned over e m b ( i d ) . One issue with this approach is that the Market1501 dataset is relatively small. When constraining the inversion to images from the training set, where identities are disjoint from the gallery, the technique struggles to generalize effectively to new identities.
Given that we have a generator at our disposal, a natural idea is to leverage it to expand the training data. We select a random identity from the latent space of identities, along with a random noise ϵ , and generate an image X i d = G e n ( ϵ , i d ) . We then proceed as previously described. However, a challenge arises from not knowing the distribution of identities in the latent space, which could be arbitrary, making it difficult to select a random identity. Our solution involves utilizing an auxiliary ID generator to learn the distribution of identities in the latent space, enabling us to sample effectively within this space.
Results of the inversion process are shown in Figure 5.
The reconstruction is satisfactory, although not flawless. Some details are lost, which could pose challenges for person re-ID.
For those interested, we also display the noise extracted from the image, which ideally contains all information not specific to the individual (middle row). The noise, assumed to have a Gaussian distribution, has been clipped within the range [−2.5, 2.5] and then renormalized to [0, 1] before visualization.
The inverted generator obtained through this process is extremely robust. In principle, it enables the derivation of the internal representation of any individual within any context, provided that the individuals and contexts are akin to those encountered in the training set.

5. Neural Network Architectures

In Section 3, it was clarified that the only trainable component within a diffusion model is the denoising network ϵ θ ( x t , α t , c ) . This network receives inputs including a noise rate α t , an image x t corrupted with a corresponding level of noise, and a condition c, which in our scenario represents the identifier of an instance. The objective is to predict the noise present in the image.
This task aligns with the conventional image-to-image paradigm, and the denoising network is typically constructed using the well-established U-net architecture [62]. The architecture of our implemented denoising network is illustrated in Figure 6; the main constituent blocks are depicted in Figure 7.
As is customary for U-nets, our network follows an encoder–decoder structure with skip connections that link corresponding layers in the encoder and decoder. This architectural choice facilitates the assimilation of both global and local structures within the image, making it highly compatible with the demands of the diffusion model.
The U-Net typically relies on parameters such as the number of downsampling blocks and the number of channels for each block, with the upsampling structure being symmetric. In our implementation, we opted for four downsampling blocks with respective channel dimensions [48, 96, 192, 384].
We emphasize that our work deals with images at a resolution of 64 × 32 ; for higher-resolution images such as 128 × 64 , an additional layer in the U-chain may be necessary.
The input pertaining to the noise variance α t is incorporated using sinusoidal positional embeddings. Subsequently, this information is vectorized, concatenated, and forwarded through the blocks along with the conditioning information of the ID and the processing of the noisy image.
Regarding the inversion networks, Img2Noise entails an image-to-image process, as the noise maintains the same spatial dimension as the input image. Based on prior studies [57], the U-Net architecture has demonstrated superior performance for such tasks.
The Img2Id model follows a similar approach, employing only the encoder portion of the U-Net. The processing is then finalized through a compact sequence of dense layers.

6. Evaluation

We conducted experiments utilizing the Market-1501 dataset [66], which comprises images of 1501 individuals captured by six cameras situated in front of an outdoor supermarket at Tsinghua University. The dataset includes 32,668 annotated bounding boxes. The training set comprises 12,936 images featuring 751 unique identities, while the gallery set contains 19,732 images encompassing 751 distinct identities. The query set consists of 3368 images, with 750 identities consistent with those in the gallery set.
The current performance of our technique falls somewhat below the state of the art. On Market1501, we achieved a mean average precision (mAP) of 73%, whereas state-of-the-art techniques attain a MAP of 96% in the supervised setting and approximately 90% in the unsupervised setting. The mAP score was computed following the guidelines provided by the dataset.
Visually, the results are satisfactory, and the errors made by the model are quite understandable given the low resolution we are working with. Distinguishing between certain identities at this resolution presents a significant challenge (see Figure 8).

7. Latent Space Exploration

In our approach, latent representations are synthesized as the features of each individual able to shape the generation of distinctive images of the person under different noises. As is typical in generative processes, this naturally leads to the creation of a highly informed and well-structured latent space, where similar persons yield similar encodings.
Figure 9 presents some examples; in the first column, we have several identities, and in the corresponding rows, we display five different individuals whose latent representations are closest, in terms of Euclidean distance, to the representation of the given identity. These images are sourced from the traditional Market1501 dataset [67], which we primarily used for our experiments. Notably, we observe the significant similarity between different identities, underscoring the non-trivial nature of the reidentification problem.
In Figure 10, we show the two closest identities, at a minimum distance of 0.14, and the two more apart from each other, at a maximum distance of 0.99. The average Euclidean distance is around 0.6.
Another intriguing operation involves deciphering how different explanatory factors of variation in the data are captured within the latent representation, akin to understanding the meaning of nucleotide sequences in a genome. However, these factors can often be entangled, meaning that a visible effect may depend on a combination of latent variables rather than a single one.
Our studies in this direction are very preliminary. As an illustration, in Figure 11, we demonstrate the effect on generation when modifying a single variable in the latent space of identities.
It is noteworthy to observe how the modification of a single variable can correspond to a significantly different identity.
Furthermore, investigating interesting semantic trajectories within the latent space can also facilitate the synthesis of new and distinct identities, which can be valuable, for instance, for data augmentation purposes.

8. Ablation and Alternatives

Several iterations of the aforementioned architecture have undergone testing.
Prior to delving into learning the distribution of latent variables using an auxiliary generator, we experimented with various regularization techniques to shape its distribution. This ranged from straightforward methods such as employing BatchNormalization layers to injecting a mild amount of noise. Subsequently, we explored a comprehensive variational model that should enforce the adoption of a Gaussian-shaped latent space through Kullback–Leibler regularization.
However, a drawback of the latter approach is the relatively weak signal influencing the definition of latent representations. This poses a challenge in balancing it with the KL component, as highlighted in [68]; in such a situation, the KL component could dominate, leading to the well-known variable collapse phenomenon [69].
The utilization of an auxiliary generator can essentially be conceptualized as a two-stage generative model [68,70]. Initially, we experimented with a variational autoencoder for implementing the auxiliary generator. However, this endeavor resulted in a notable increase in reconstruction loss, which corresponded to a loss of variance in generated data [71]. In contrast, a diffusion model appears to learn the distribution much more effectively in this regard.
Exploring a different research direction, we sought to increase the separation between distinct latent representations through the incorporation of a "repulsive" loss, which was proportionate to the inverse of their distance in the space, akin to a form of magnetic repulsion. This was complemented by a weak attractive loss, which encouraged points to remain close to the origin, effectively centering the latent space around it.
While we retained the latter, the repulsive loss did not appear to enhance performance and could potentially be detrimental. Our hypothesis is that the loss primarily impacts latent variables with minimal significance, attempting to segregate them across different subjects. This segregation is artificial and lacks semantic predictability, rendering it unreconstructable from the input images.
Finally, we conducted several experiments on the Embedding module for identities. In the current implementation, it comprises a single Keras Embedding layer. Initially, this layer expands labels to their categorical encoding and subsequently maps them to a latent space of the desired dimension through a single dense layer.
In principle, this setup should not impose any restrictions, as any embedding can potentially be learned from a categorical description. However, what remains unclear is the model’s flexibility to adapt and enhance encodings during training, along with the acquisition of additional knowledge. To explore this aspect, we experimented with more complex embedding modules, although they did not yield any discernible benefits.

9. Conclusions

The work presented in this article is first of all a divertissment; we aimed to test the conditional capabilities of generative diffusion models in a complex scenario. Specifically, we sought to learn the latent representation of different individuals as the shared information necessary to condition the generation of images of the given person from varying noise. This approach enables the separation of an individual’s identity from other instance-specific information, such as pose and background, expressed as part of the noise.
There are numerous avenues for improvement in this work. Both the generator and the network used for its inversion could be enhanced. A natural research direction involves repeating the experiments with images at the original 128 × 68 spatial resolution, potentially expanding the dimension of the latent space. Additionally, a detailed analysis of the errors made during the re-identification process could lead to valuable insights and potential enhancements. Leveraging the generator for intelligent and targeted data augmentation presents another intriguing possibility.
A limitation of our approach is its reliance on supervised training; to train the conditional generator, a significant number of instances of the same identity in different poses and contexts are required. Exploring whether the methodology can be extended to an unsupervised scenario presents a complex and interesting challenge. It is plausible that typical unsupervised person re-ID approaches, predominantly based on clustering, centroids, and pseudo-labels [72,73,74], could be adapted to our approach.

Author Contributions

Conceptualization, A.A.; Methodology, A.A.; Software, A.A., S.F. and L.O.; Validation, L.O.; Formal analysis, S.F.; Investigation, S.F. and L.O.; Resources, S.F.; Data curation, L.O.; Writing—original draft, A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Future AI Research (FAIR) project of the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 1.3 funded from the European Union—NextGenerationEU.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The application described in this paper is open-source. The software can be downloaded from the following github repository: https://github.com/asperti/GenerativePersonReID (accessed on 1 February 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bukhari, M.; Yasmin, S.; Naz, S.; Maqsood, M.; Rew, J.; Rho, S. Language and vision based person re-identification for surveillance systems using deep learning with LIP layers. Image Vis. Comput. 2023, 132, 104658. [Google Scholar] [CrossRef]
  2. Kim, K.; Kim, M.J.; Kim, H.; Park, S.; Paik, J. Person Re-identification Method Using Text Description Through CLIP. In Proceedings of the 2023 International Conference on Electronics, Information, and Communication (ICEIC), Singapore, 5–8 February 2023; pp. 1–4. [Google Scholar] [CrossRef]
  3. Ming, Z.; Zhu, M.; Wang, X.; Zhu, J.; Cheng, J.; Gao, C.; Yang, Y.; Wei, X. Deep learning-based person re-identification methods: A survey and outlook of recent works. Image Vis. Comput. 2022, 119, 104394. [Google Scholar] [CrossRef]
  4. Chen, J.; Wang, Y.; Tang, Y.Y. Person Re-identification by Exploiting Spatio-Temporal Cues and Multi-view Metric Learning. IEEE Signal Process. Lett. 2016, 23, 998–1002. [Google Scholar] [CrossRef]
  5. Chung, D.; Tahboub, K.; Delp, E.J. A Two Stream Siamese Convolutional Neural Network for Person Re-identification. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1992–2000. [Google Scholar] [CrossRef]
  6. Wei, L.; Zhang, S.; Gao, W.; Tian, Q. Person Transfer GAN to Bridge Domain Gap for Person Re-Identification 2018. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 79–88. [Google Scholar] [CrossRef]
  7. Liu, X.; Tan, H.; Tong, X.; Cao, J.; Zhou, J. Feature preserving GAN and multi-scale feature enhancement for domain adaption person Re-identification. Neurocomputing 2019, 364, 108–118. [Google Scholar] [CrossRef]
  8. Li, Y.; Chen, S.; Qi, G.; Zhu, Z.; Haner, M.; Cai, R. A GAN-Based Self-Training Framework for Unsupervised Domain Adaptive Person Re-Identification. J. Imaging 2021, 7, 62. [Google Scholar] [CrossRef] [PubMed]
  9. Tang, G.; Gao, X.; Chen, Z.; Zhong, H. Unsupervised adversarial domain adaptation with similarity diffusion for person re-identification. Neurocomputing 2021, 442, 337–347. [Google Scholar] [CrossRef]
  10. Li, Y.; He, J.; Zhang, T.; Liu, X.; Zhang, Y.; Wu, F. Diverse part discovery: Occluded person re-identification with part-aware transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2898–2907. [Google Scholar]
  11. Cao, G.; Jo, K.H. Unsupervised Person Re-Identification with Transformer-based Network for Intelligent Surveillance Systems. In Proceedings of the 2021 IEEE 30th International Symposium on Industrial Electronics (ISIE), Kyoto, Japan, 20–23 June 2021; pp. 1–6. [Google Scholar]
  12. Chen, Y.; Xia, S.; Zhao, J.; Zhou, Y.; Niu, Q.; Yao, R.; Zhu, D.; Liu, D. ResT-ReID: Transformer block-based residual learning for person re-identification. Pattern Recognit. Lett. 2022, 157, 90–96. [Google Scholar] [CrossRef]
  13. Chen, X.; Xie, S.; He, K. An empirical study of training self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9620–9629. [Google Scholar]
  14. Perwaiz, N.; Shahzad, M.; Fraz, M. Ubiquitous vision of transformers for person re-identification. Mach. Vis. Appl. 2023, 34, 27. [Google Scholar] [CrossRef]
  15. Zhou, S.; Wang, F.; Huang, Z.; Wang, J. Discriminative Feature Learning With Consistent Attention Regularization for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8039–8048. [Google Scholar] [CrossRef]
  16. Huang, Y.; Lian, S.; Hu, H.; Chen, D.; Su, T. Multiscale Omnibearing Attention Networks for Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1790–1803. [Google Scholar] [CrossRef]
  17. Huang, Y.; Peng, P.; Jin, Y.; Li, Y.; Xing, J. Domain Adaptive Attention Learning for Unsupervised Person Re-Identification. In Proceedings of the The Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11069–11076. [Google Scholar] [CrossRef]
  18. Saber, S.; Meshoul, S.; Amin, K.; Pławiak, P.; Hammad, M. A Multi-Attention Approach for Person Re-Identification Using Deep Learning. Sensors 2023, 23, 3678. [Google Scholar] [CrossRef]
  19. Somers, V.; De Vleeschouwer, C.; Alahi, A. Body Part-Based Representation Learning for Occluded Person Re-Identification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 1613–1623. [Google Scholar]
  20. Wu, J.J.; Chang, K.H.; Lin, I.C. Generalizable person re-identification with part-based multi-scale network. Multimed. Tools Appl. 2023, 82, 38639–38666. [Google Scholar] [CrossRef]
  21. Fu, D.; Chen, D.; Bao, J.; Yang, H.; Yuan, L.; Zhang, L.; Li, H.; Chen, D. Unsupervised Pre-Training for Person Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 14750–14759. [Google Scholar] [CrossRef]
  22. Yang, Z.; Jin, X.; Zheng, K.; Zhao, F. Unleashing Potential of Unsupervised Pre-Training with Intra-Identity Regularization for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 14278–14287. [Google Scholar] [CrossRef]
  23. Chen, W.; Xu, X.; Jia, J.; Luo, H.; Wang, Y.; Wang, F.; Jin, R.; Sun, X. Beyond Appearance: A Semantic Controllable Self-Supervised Learning Framework for Human-Centric Visual Tasks 2023. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 15050–15061. [Google Scholar] [CrossRef]
  24. Le-Khac, P.H.; Healy, G.; Smeaton, A.F. Contrastive Representation Learning: A Framework and Review. IEEE Access 2020, 8, 193907–193934. [Google Scholar] [CrossRef]
  25. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality Reduction by Learning an Invariant Mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1735–1742. [Google Scholar] [CrossRef]
  26. Wang, M.; Lai, B.; Huang, J.; Gong, X.; Hua, X.S. Graph-Induced Contrastive Learning for Intra-Camera Supervised Person Re-Identification. IEEE Access 2021, 9, 20850–20860. [Google Scholar] [CrossRef]
  27. Hu, S.; Zhang, X.; Xie, X. Decoupled Contrastive Learning for Intra-Camera Supervised Person Re-identification. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 2628–2665. [Google Scholar] [CrossRef]
  28. Shi, X.; Liu, H.; Shi, W.; Zhou, Z.; Li, Y. Boosting Person Re-Identification with Viewpoint Contrastive Learning and Adversarial Training. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  29. Wang, J.; Song, Y.; Leung, T.; Rosenberg, C.; Wang, J.; Philbin, J.; Chen, B.; Wu, Y. Learning Fine-Grained Image Similarity with Deep Ranking. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1386–1393. [Google Scholar] [CrossRef]
  30. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  31. Hermans, A.; Beyer, L.; Leibe, B. In defense of the triplet loss for person re-identification. arXiv 2017, arXiv:1703.07737. [Google Scholar]
  32. Yuan, Y.; Chen, W.; Yang, Y.; Wang, Z. In Defense of the Triplet Loss Again: Learning Robust Person Re-Identification with Fast Approximated Triplet Loss and Label Distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  33. Zhang, S.; Zhang, Q.; Wei, X.; Zhang, Y.; Xia, Y. Person Re-Identification With Triplet Focal Loss. IEEE Access 2018, 6, 78092–78099. [Google Scholar] [CrossRef]
  34. Si, T.; Zhang, Z.; Liu, S. Compact Triplet Loss for person re-identification in camera sensor networks. Ad Hoc Netw. 2019, 95, 101984. [Google Scholar] [CrossRef]
  35. Wieczorek, M.; Rychalska, B.; Dąbrowski, J. On the unreasonable effectiveness of centroids in image retrieval. In Proceedings of the 28th International Conference on Neural Information Processing, ICONIP 2021, Sanur, Bali, Indonesia, 8–12 December 2021; pp. 212–223. [Google Scholar]
  36. Zhao, D.; Chen, C.; Li, D. Multi-stage attention and center triplet loss for person re-identication. Appl. Intell. 2022, 52, 3077–3089. [Google Scholar] [CrossRef]
  37. Alnissany, A.; Dayoub, Y. Modified centroid triplet loss for person re-identification. J. Big Data 2023, 10, 74. [Google Scholar] [CrossRef]
  38. Chen, W.; Chen, X.; Zhang, J.; Huang, K. Beyond Triplet Loss: A Deep Quadruplet Network for Person Re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1320–1329. [Google Scholar] [CrossRef]
  39. Sohn, K. Improved deep metric learning with multi-class n-pair loss objective. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  40. Horiguchi, S.; Ikami, D.; Aizawa, K. Significance of Softmax-based Features in Comparison to Distance Metric Learning-based Features. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1279–1285. [Google Scholar] [CrossRef]
  41. Zhai, Y.; Guo, X.; Lu, Y.; Li, H. In Defense of the Classification Loss for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  42. Ni, X.; Rahtu, E. Flipreid: Closing the gap between training and inference in person re-identification. In Proceedings of the 9th European Workshop on Visual Information Processing, EUVIP, Paris, France, 23–25 June 2021; pp. 1–6. [Google Scholar] [CrossRef]
  43. Chen, H.; Wang, Y.; Lagadec, B.; Dantcheva, A.; Bremond, F. Joint generative and contrastive learning for unsupervised person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2004–2013. [Google Scholar]
  44. Varga, D.; Szirányi, T. Person re-identification based on deep multi-instance learning. In Proceedings of the 2017 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 1559–1563. [Google Scholar]
  45. Song, C.; Huang, Y.; Ouyang, W.; Wang, L. Mask-guided contrastive attention model for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1179–1188. [Google Scholar]
  46. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020. [Google Scholar]
  47. Song, J.; Meng, C.; Ermon, S. Denoising Diffusion Implicit Models. In Proceedings of the 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
  48. Deng, W.; Zheng, L.; Ye, Q.; Kang, G.; Yang, Y.; Jiao, J. Image-Image Domain Adaptation with Preserved Self-Similarity and Domain-Dissimilarity for Person Re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 994–1003. [Google Scholar] [CrossRef]
  49. Yanbei, C.; Zhu, X.; Gong, S. Instance-Guided Context Rendering for Cross-Domain Person Re-Identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  50. Verma, A.; Subramanyam, A.; Wang, Z.; Satoh, S.; Shah, R.R. Unsupervised Domain Adaptation for Person Re-Identification Via Individual-Preserving and Environmental-Switching Cyclic Generation. IEEE Trans. Multimed. 2021, 25, 364–377. [Google Scholar] [CrossRef]
  51. Zhu, Y.; Deng, C.; Cao, H.; Wang, H. Object and background disentanglement for unsupervised cross-domain person re-identification. Neurocomputing 2020, 403, 88–97. [Google Scholar] [CrossRef]
  52. Dhariwal, P.; Nichol, A.Q. Diffusion Models Beat GANs on Image Synthesis. In Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, Virtual, 6–14 December 2021; pp. 8780–8794. [Google Scholar]
  53. Ramesh, A.; Dhariwal, P.; Nichol, A.; Chu, C.; Chen, M. Hierarchical Text-Conditional Image Generation with CLIP Latents. arXiv 2022, arXiv:2204.06125. [Google Scholar]
  54. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.; Ghasemipour, S.K.S.; Ayan, B.K.; Mahdavi, S.S.; Lopes, R.G.; et al. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. In Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, 28 November–9 December 2022; Available online: http://papers.nips.cc/paper_files/paper/2022/hash/ec795aeadae0b7d230fa35cbaf04c041-Abstract-Conference.html (accessed on 20 December 2023).
  55. Ho, J.; Salimans, T.; Gritsenko, A.; Chan, W.; Norouzi, M.; Fleet, D.J. Video diffusion models. arXiv 2022, arXiv:2204.03458. [Google Scholar]
  56. Asperti, A.; Merizzi, F.; Paparella, A.; Pedrazzi, G.; Angelinelli, M.; Colamonaco, S. Precipitation nowcasting with generative diffusion models. arXiv 2023, arXiv:2308.06733. [Google Scholar]
  57. Asperti, A.; Evangelista, D.; Marro, S.; Merizzi, F. Image Embedding for Denoising Generative Models. arXiv 2022, arXiv:2301.07485. [Google Scholar] [CrossRef]
  58. Asperti, A.; Colasuonno, G.; Guerra, A. Portrait Reification with Generative Diffusion Models. Appl. Sci. 2023, 13, 6487. [Google Scholar] [CrossRef]
  59. Asperti, A.; Colasuonno, G.; Guerra, A. Head Rotation in Denoising Diffusion Models. arXiv 2023, arXiv:2308.06057. [Google Scholar]
  60. Ho, J.; Salimans, T. Classifier-Free Diffusion Guidance. arXiv 2022, arXiv:2207.12598. [Google Scholar] [CrossRef]
  61. Odena, A.; Olah, C.; Shlens, J. Conditional Image Synthesis with Auxiliary Classifier GANs. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6–11 August 2017; Volume 70, pp. 2642–2651. [Google Scholar]
  62. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  63. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  64. Kingma, D.; Salimans, T.; Poole, B.; Ho, J. Variational diffusion models. Adv. Neural Inf. Process. Syst. 2021, 34, 21696–21707. [Google Scholar]
  65. Nichol, A.Q.; Dhariwal, P. Improved denoising diffusion probabilistic models. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8162–8171. [Google Scholar]
  66. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable Person Re-identification: A Benchmark. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar] [CrossRef]
  67. Chen, M.; Wang, Z.; Zheng, F. Benchmarks for Corruption Invariant Person Re-identification. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks 1, NeurIPS Datasets and Benchmarks 2021, Virtual, 6–14 December 2021; Available online: https://datasets-benchmarks-proceedings.neurips.cc/paper_files/paper/2021/file/f457c545a9ded88f18ecee47145a72c0-Paper-round2.pdf (accessed on 20 December 2023).
  68. Asperti, A.; Trentin, M. Balancing Reconstruction Error and Kullback-Leibler Divergence in Variational Autoencoders. IEEE Access 2020, 8, 199440–199448. [Google Scholar] [CrossRef]
  69. Asperti, A.; Evangelista, D.; Piccolomini, E.L. A Survey on Variational Autoencoders from a Green AI Perspective. SN Comput. Sci. 2021, 2, 301. [Google Scholar] [CrossRef]
  70. Dai, B.; Wipf, D.P. Diagnosing and enhancing VAE models. In Proceedings of the Seventh International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  71. Asperti, A. Variance Loss in Variational Autoencoders. In Proceedings of the Machine Learning, Optimization, and Data Science—6th International Conference, LOD 2020, Siena, Italy, 19–23 July 2020. [Google Scholar]
  72. Fan, H.; Zheng, L.; Yang, Y. Unsupervised Person Re-identification: Clustering and Fine-tuning. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2017, 14, 1–18. [Google Scholar] [CrossRef]
  73. Wang, S.; Zhang, L.; Chen, W.; Wang, F.; Li, H. Refining pseudo labels for unsupervised Domain Adaptive Re-Identification. Knowl.-Based Syst. 2021, 242, 108336. [Google Scholar] [CrossRef]
  74. Yan, T.; Zhu, K.; Guo, H.; Zhu, G.; Tang, M.; Wang, J. Plug-and-Play Pseudo Label Correction Network for Unsupervised Person Re-identification. arXiv 2022, arXiv:2206.06607. [Google Scholar] [CrossRef]
Figure 1. Conditional generative model (a) and its inversion (b). We trained a diffusion generative model conditioned on an embedding of people’s identity. The embedding of identity labels was learned along with the generator. In reverting the conditional generator, we obtained a model that was able to split the image into the identity of the person, and all of the contextual information (pose, background, and so on) that was mapped back to the noise.
Figure 1. Conditional generative model (a) and its inversion (b). We trained a diffusion generative model conditioned on an embedding of people’s identity. The embedding of identity labels was learned along with the generator. In reverting the conditional generator, we obtained a model that was able to split the image into the identity of the person, and all of the contextual information (pose, background, and so on) that was mapped back to the noise.
Sensors 24 01240 g001
Figure 2. Conditional generation. On the left, three sample identities taken from the Market1501 dataset. We sample 5 random noises, and for each of them we generate an image conditioned by the given identity; the noise in each column is always the same.
Figure 2. Conditional generation. On the left, three sample identities taken from the Market1501 dataset. We sample 5 random noises, and for each of them we generate an image conditioned by the given identity; the noise in each column is always the same.
Sensors 24 01240 g002
Figure 3. Direct and reverse diffusion. Picture from [56].
Figure 3. Direct and reverse diffusion. Picture from [56].
Sensors 24 01240 g003
Figure 4. The inversion of the generator can be realized through a single network (a), or split in two distinct networks (b).
Figure 4. The inversion of the generator can be realized through a single network (a), or split in two distinct networks (b).
Sensors 24 01240 g004
Figure 5. Reconstruction examples. In the fist row, we have original gallery images, in the middle row the noise synthesized by inversion, and in the third row the reconstructed images.
Figure 5. Reconstruction examples. In the fist row, we have original gallery images, in the middle row the noise synthesized by inversion, and in the third row the reconstructed images.
Sensors 24 01240 g005
Figure 6. Denoising network architecture.
Figure 6. Denoising network architecture.
Sensors 24 01240 g006
Figure 7. Main blocks of the Denoising Network. (a): The input ID undergoes processing through an embedding layer; the resulting output, after reshaping and upsampling, is concatenated with the various downblocks. (b): The nth downblock receives two inputs: the embedded ID and the output of the previous block. The block outputs two elements: the skipping connections for the upblock at the same level and the output for the next block. (c): a traditional Residual Block.
Figure 7. Main blocks of the Denoising Network. (a): The input ID undergoes processing through an embedding layer; the resulting output, after reshaping and upsampling, is concatenated with the various downblocks. (b): The nth downblock receives two inputs: the embedded ID and the output of the previous block. The block outputs two elements: the skipping connections for the upblock at the same level and the output for the next block. (c): a traditional Residual Block.
Sensors 24 01240 g007
Figure 8. Person re-identification. In each row, the first image is the query, and the successive images are the 5 best matchings in the gallery.
Figure 8. Person re-identification. In each row, the first image is the query, and the successive images are the 5 best matchings in the gallery.
Sensors 24 01240 g008
Figure 9. Inherent clustering of the latent space. For every ID in the initial column, we display five distinct individuals whose latent representations are closer, in terms of Euclidean distance, to the representation of the specified ID.
Figure 9. Inherent clustering of the latent space. For every ID in the initial column, we display five distinct individuals whose latent representations are closer, in terms of Euclidean distance, to the representation of the specified ID.
Sensors 24 01240 g009
Figure 10. On the left, we have the two closest identities in the latent space of Market1501; on the right, the two more distant ones.
Figure 10. On the left, we have the two closest identities in the latent space of Market1501; on the right, the two more distant ones.
Sensors 24 01240 g010
Figure 11. Effect of latent variables on generation. In each row, we keep the same random noise and vary a given variable in a predefined range. The specific variable and its value are indicated above the generated image.
Figure 11. Effect of latent variables on generation. In each row, we keep the same random noise and vary a given variable in a predefined range. The specific variable and its value are indicated above the generated image.
Sensors 24 01240 g011
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Asperti, A.; Fiorilla, S.; Orsini, L. A Generative Approach to Person Reidentification. Sensors 2024, 24, 1240. https://doi.org/10.3390/s24041240

AMA Style

Asperti A, Fiorilla S, Orsini L. A Generative Approach to Person Reidentification. Sensors. 2024; 24(4):1240. https://doi.org/10.3390/s24041240

Chicago/Turabian Style

Asperti, Andrea, Salvatore Fiorilla, and Lorenzo Orsini. 2024. "A Generative Approach to Person Reidentification" Sensors 24, no. 4: 1240. https://doi.org/10.3390/s24041240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop