Erasing, Transforming, and Noising Defense Network for Occluded Person Re-Identification
Abstract
Occlusion perturbation presents a significant challenge in person re-identification (re-ID), and existing methods that rely on external visual cues require additional computational resources and only consider the issue of missing information caused by occlusion. In this paper, we propose a simple yet effective framework, termed Erasing, Transforming, and Noising Defense Network (ETNDNet), which treats occlusion as a noise disturbance and solves occluded person re-ID from the perspective of adversarial defense. In the proposed ETNDNet, we introduce three strategies: Firstly, we randomly erase the feature map to create an adversarial representation with incomplete information, enabling adversarial learning of identity loss to protect the re-ID system from the disturbance of missing information. Secondly, we introduce random transformations to simulate the position misalignment caused by occlusion, training the extractor and classifier adversarially to learn robust representations immune to misaligned information. Thirdly, we perturb the feature map with random values to address noisy information introduced by obstacles and non-target pedestrians, and employ adversarial gaming in the re-ID system to enhance its resistance to occlusion noise. Without bells and whistles, ETNDNet has three key highlights: (i) it does not require any external modules with parameters, (ii) it effectively handles various issues caused by occlusion from obstacles and non-target pedestrians, and (iii) it designs the first GAN-based adversarial defense paradigm for occluded person re-ID. Extensive experiments on six public datasets fully demonstrate the effectiveness, superiority, and practicality of the proposed ETNDNet. The code will be released at https://github.com/nengdong96/ETNDNet.
Index Terms:
Occlusion perturbation, adversarial representation, GAN-based adversarial defense.I Introduction
Person re-identification (re-ID) aims at matching the same pedestrians from different cameras, providing technical support for a series of intelligent security tasks such as pedestrian tracking, autonomous driving, and activity recognition [1, 2, 3]. With the rapid development of the re-ID community, various superior algorithms have emerged [4, 5, 6, 7, 8]. However, most of these methods ideally assume that the whole body of each pedestrian is visible, which is challenging to satisfy in real scenarios due to the inevitable occlusion.


Empirically, occlusion mainly causes the following issues [9]: i). Missing Information. Obstacles such as warning signs can cause the loss of identity-related information, such as legs and even the torso. ii). Position Misalignment. In case of occlusion, the detected pedestrian box only contains a part of the human body. Obviously, position misalignment occurs when holistic and partial images undergo the same scaling. iii). Noisy Information. Besides the loss of valuable information, obstacles and non-target pedestrians introduce irrelevant noisy information. Affected by these issues, the model tends to focus only on locally visible regions in occluded images (neglecting obscured and misaligned parts) and may even pay attention to noise areas associated with obstacles (see Figure 1(a)). Existing methods [10, 11, 12] often rely on external tools such as pose key-points [13] and human semantics [14] to detect visible pedestrian regions for identity matching. While effective, this approach requires additional computation costs and only considers the issue of missing information. Moreover, the visual cues produced by external tools may be inaccurate in complex scenes (as shown in Figure 1(b)). Consequently, accurately locating the visible areas of the target pedestrian becomes challenging for the model, leading to errors in pedestrian identity matching.
Adversarial defense methods have gained significant attention due to their ability to protect Deep Neural Networks (DNNs) from noise perturbations [15, 16, 17]. Considering that occlusion can be regarded as a kind of noise disturbance to re-ID models, it is plausible to protect re-ID models immune to occlusion perturbation from the perspective of adversarial defense. Previous research has attempted to solve the problem of occlusion from this perspective. For example, Huang et al. [18] generated adversarial occluded samples by substituting a region of the original images with all zeros, subsequently fine-tuning the re-ID model to enhance the robustness of learned features. Similarly, Zhao et al. [19] proposed an easy-to-hard erasing approach to generate aggressive occluded data, which encourages the network to pay attention to non-occluded regions. However, these methods rely on generating adversarial samples to endow re-ID models with the ability to defend against occlusion interference. On the one hand, this way of generating adversarial samples increases the training cost due to the expansion of the sample base. On the other hand, the generated data can only simulate the issue of missing information while ignoring position misalignment and noisy information. Additionally, adversarial samples obtained through perturbation on the original inputs are weakly aggressive, while strongly aggressive adversarial examples are crucial for training models with robust defense capabilities. Moreover, the aforementioned methods focus on the information in non-erased areas without enhancing its discriminability, which limits performance improvement.
According to the findings in [20], the main culprit of the vulnerability of DNNs is the adversarial disturbance acting on sensitive features. Therefore, it is more effective to directly apply various perturbations to pedestrian feature maps, which not only eliminate the need for regenerative data but can also be more aggressive than adversarial samples since feature maps inherently contain rich semantic information. To distinguish from the above adversarial samples, we refer to such adversarial examples formed by perturbing feature maps as adversarial representations. In addition, drawing inspiration from Generative Adversarial Networks (GAN) [21], which use a generator to produce samples and a discriminator to distinguish between real and fake samples, we introduce a similar paradigm for occluded person re-ID. For the person re-ID task, the proposed paradigm mainly includes an extractor for learning pedestrian representations and a classifier for identification. While the classifier can accurately identify benign representations, it fails to classify disturbed representations. However, if the representations learned by the extractor are robust enough, even when affected by occlusion disturbances, the classifier can correctly identify the adversarial representations as it does for benign ones. Therefore, it is desirable to endow the re-ID system with the ability to defend against occlusion perturbation through an adversarial game between the feature extractor and identity classifier. Notably, our proposed approach, unlike the previous GAN-based adversarial defense paradigm [22], is specifically tailored for person re-ID, an image retrieval task, and does not require learning to generate adversarial perturbations for original inputs.
Motivated by the above discussion, we propose an Erasing, Transforming, and Noising Defense Network (ETNDNet) to protect the re-ID system against occlusion perturbations. Specifically, to address the issue of missing information, we employ random erasing on the feature map to generate an adversarial representation, which maximizes the identity loss to optimize the classifier and simultaneously minimizes the loss to optimize the feature extractor. By continuously confronting each other, the feature extractor is driven to learn a robust representation that deceives the classifier into correctly identifying the attacked version, which indicates that the learned representation is immune to information loss. Moreover, we randomly select an area on the feature map and transform its pixels to another area, simulating position misalignment. Similarly, training the classifier enables it to identify the clean feature map accurately, but fails to classify the perturbed one. And in turn, training the feature extractor enables the classifier to assign the correct identity category to the adversarial representation, thereby ensuring that the feature map remains unaffected by misaligned information. Recognition performance is negatively impacted by noisy information from obstacles and non-target pedestrians, yet limited studies have attempted to overcome this challenge. To this end, we further apply random noise to the feature map and accordingly train the re-ID model in an adversarial way, enhancing the robustness of feature maps against noisy information. Importantly, the above paradigm does not require the introduction of redundant components, offering significant advantages in terms of model simplicity.
The main contributions of this paper can be summarized as follows:
-
•
Without external auxiliary models and additional trainable modules, we develop a simple yet effective adversarial defense method to learn occlusion-robust pedestrian representations, in which the first GAN-based adversarial defense paradigm is designed for person re-ID tasks.
-
•
Our proposed ETNDNet endows the re-ID system with the ability to handle occlusion perturbations, alleviating the adverse effects of missing information and position misalignment on recognition performance. Particularly, our approach effectively addresses the problem of noisy information resulting from occlusions caused by obstacles and non-target pedestrians, which has often been overlooked in previous studies.
-
•
We conduct extensive experiments on six public datasets and compare the performance with existing state-of-the-art methods to demonstrate the superiority of the proposed method. Moreover, we evaluate the effectiveness and practicality of our ETNDNet through various aspects including ablation studies and analysis of model complexity.
The remainder of this paper is organized as follows: The related works are introduced in Section II; The details of the proposed method are elaborated in Section III; The experimental comparison and analysis are reflected in Section IV; The conclusion of the present paper is summarized in Section V.
II Related Work
II-A Holistic Person Re-ID
Most existing re-ID studies assume that the pedestrian body is holistic and can be used to learn a complete representation. Hand-crafted feature-based methods [23, 24, 25] adopt pre-designed descriptors to capture pedestrian characteristics. For example, Farenzena et al. [23] computed maximally stable color regions and recurrent high-structured patches to extract color information and texture details, respectively. Liao et al. [24] proposed a local maximal occurrence algorithm that models salient representations to handle viewpoint and illumination changes. Moreover, Li et al.[25] combined Gaussian of Gaussian (GOG) features with dictionary learning to reduce domain discrepancies across datasets. However, with the expansion of data, the applicability of such methods becomes unsatisfactory since they fail to exploit sample distribution adequately. To overcome these limitations, deep learning-based methods [26, 27, 28, 29, 30] have emerged. For example, Yi et al. [26] first introduced deep neural networks (DNNs) for person re-ID and built a siamese model to capture pedestrian similarities. Considering the importance of fine-grained information in re-ID, Sun et al. [27] developed a part-aware feature extraction method with an adaptive partitioning strategy to align body parts. Li et al.[31] exploited features from multiple camera views to enable multi-view imaginative reasoning. Recently, additional techniques such as BNNeck [4] and GeM [5] have been developed, further advancing person re-ID tasks. Nevertheless, in crowded scenes where pedestrians are often occluded by obstacles or other persons, these methods struggle to achieve satisfactory performance. To address this, our work focuses on solving the occlusion problem in person re-ID while achieving excellent recognition rates in holistic scenarios.
II-B Occluded Person Re-ID
Compared with holistic re-ID, occluded person re-ID is more practical and challenging, where the probe samples are occluded and the gallery database contains both holistic and occluded images. To encourage the model to focus on human body parts, Zhuo et al. [32] simulated occluded samples of holistic data using an occlusion simulator and introduced a co-saliency branch to assist the feature extractor in capturing identity-related information. Moreover, Kiran et al. [33] adopted a teacher-student manner to match the distributions of intra- and inter-class distances of occluded and holistic data, thereby separating visible regions from occluded images. To extract discriminative features of visible regions, He et al. [34] inserted a foreground-aware pyramid reconstruction module that adaptively assigns weights to body and occlusion parts. Despite their effectiveness, learning occlusion-robust representations remains a significant challenge for these methods. Our proposed ETNDNet falls into the external model-free category, however, it endows the re-ID model with the intrinsic ability to defend against occlusion perturbations.
Recent studies have focused on guiding the re-ID model to extract features from the visible region by utilizing external tools. For example, Miao et al.[35] integrated a pre-trained pose estimator into the re-ID model, generating pose-aware heatmaps that indicate if specific body parts are occluded. Similarly, Gao et al. [10] developed a pose-guided attention mechanism and a visibility predictor to end-to-end match identities. To capture high-order relation information of human parts, Wang et al. [11] treated local features learned by keypoints as graph nodes to model the topology structure of the human body. Leveraging the intrinsic connection between re-ID and human parsing, Zhang et al. [36] introduced a semantic branch to extract global and local features simultaneously. Moreover, Zheng et al. [37] proposed a pose-guided feature learning approach with knowledge distillation to obtain semantics-aligned representation. Hou et al. [12] performed person foreground segmentation and devised a spatial region feature completion module to reason about occluded areas. Nevertheless, the performance of these methods largely relies on the accuracy of pose landmarks and body parsing. Additionally, the use of the auxiliary model increases computational costs and may compromise the accuracy of visible cues. In contrast, our method overcomes these limitations and achieves state-of-the-art performance.
II-C Adversarial Attack and Defense
The application of DNNs has led to significant advancements in various computer vision tasks [38, 39, 40, 41]. However, Szegedy et al. [42] discovered that even highly performing models may produce incorrect inferences when imperceptible perturbations are applied to images. Building on this observation, Goodfellow et al. [15] introduced adversarial examples by adding noise to original inputs using the fast gradient sign method. Kurakin et al. [43] proposed a multi-step attack method that iteratively classifies attacked samples into categories with the lowest classification probabilities. For person re-ID tasks, recent studies [44, 45] have demonstrated that the re-ID system is also vulnerable to adversarial attacks, such as images of pedestrians wearing different clothes or captured by different cameras, which can lead to misidentification of individuals. Specifically, Wang et al. [44] designed the advPattern algorithm to explore the impact of clothing changes on recognition performance. To maintain visual quality in adversarial examples, Wang et al. [45] employed differentiable multi-shot sampling to control the number of malicious pixels and proposed a novel perception loss to ensure inconspicuous attacks.
To mitigate the interference of noise to DNNs, the concept of adversarial defense has been introduced. Specifically, Zheng et al. [16] proposed joint fine-tuning of pre-trained models using clean samples and their corresponding distorted copies to improve the accuracy of predicting adversarial examples. Lyu et al. [17] developed a family of gradient regularization techniques to constrain the optimization of DNNs and defend against perturbations. Gao et al. [46] enhanced model robustness by removing irrelevant features through a mask layer positioned between the feature extractor and classifier. Building on these studies, Wang et al. [47] proposed a multi-expert adversarial attack detection approach to differentiate perturbed examples for person re-ID tasks. Treating occlusion as a form of noise perturbation, Huang et al. [18] and Zhao et al. [19] drove models to extract robust features using brute-force adversarial defense that requires perturbed examples. In contrast, our method generates reasonable adversarial representations to simulate various occlusion perturbations and trains the feature extractor and identity classifier using a novel GAN-based adversarial defense paradigm, effectively defending the re-ID system against occlusion interference.

III The proposed methods
As illustrated in Figure 2, the proposed ETNDNet consists of a baseline framework and an occlusion defense system. The former is a standard re-ID model, which is trained to ensure the discriminativeness of learned features. The latter includes erasing defense, transforming defense, and noising defense modules. In these modules, we perturb the feature maps in different manners to generate multiple adversarial representations that can simulate various issues caused by occlusion, and accordingly train the feature extractor and identity classifier in an adversarial way to endow the re-ID model with the ability to defend against occlusion perturbation, thereby promoting the occlusion robustness of the learned features. In the following, we will describe the key technologies of each part in detail.
III-A Baseline Framework
We adopt the standard re-ID model as our baseline framework, which is commonly employed for the occluded person re-ID task. This model contains a feature extractor with ResNet-50 [48] pre-trained on ImageNet [49] and an identity classifier composed of a pooling layer, a normalization layer, and a fully-connected layer. As illustrated in Figure 2, given a pedestrian image , the feature extractor is responsible for learning representation, and the classifier is utilized to predict which identity the person belongs to. Let and respectively represent the feature extractor and identity classifier, we deploy the cross-entropy loss to optimize them:
(1) |
where is the number of pedestrians, denotes the prediction logits of class , is associated with the ground truth :
(2) |
where is a constant and set to be 0.1. The association is derived from the label smoothing (LS) [50] technique, which effectively avoids the risk of overfitting.
III-B Adversarial Representations
The above baseline branch ensures the basic discriminability of learned features. However, the model is vulnerable to occlusion interference. As mentioned before, it is feasible to train an occlusion-robust network from the perspective of adversarial defense. In this process, generating reasonable adversarial examples to simulate occlusion issues is crucial. The adversarial samples generated in [18, 19] can achieve this purpose. However, they are less aggressive and can only simulate the issue of missing information. To this end, we apply perturbation on pedestrian feature maps by erasing, transforming, and noising to generate multiple adversarial representations to simulate various occlusion issues.
Erasing. We randomly erase the feature map to acquire an adversarial representation with incomplete information, which can simulate the issue of pedestrian identity information loss caused by occlusion. Suppose the feature map is , where , , and denote the height, width, and the number of channels, respectively. Similar to [51], we first compute the spatial size of the feature map. Next, we randomly initialize a proportion to determine the size of the erasing rectangle region , where . And then, we arbitrarily take an aspect ration and accordingly compute the height and width of , where . Finally, we randomly select a point in . If and , can be determined. Otherwise, the above process is repeated until an appropriate is selected. Let each pixel belonging to be 0, and the adversarial representation can be formulated as:
(3) |
Transforming. Occlusion may cause the pedestrian information at a certain position on the non-occluded image to appear at another position on the occluded image. Therefore, we randomly transform the feature map to simulate such a position misalignment issue, which is an ability that the adversarial samples do not possess. Firstly, we randomly initialize a proportion and an aspect ratio to determine the height and width of areas to be transformed, where and . Next, arbitrarily choosing two points , and , if , , , and , two areas and can be located. Accordingly, we transform the pixels of to for generating an adversarial representation with misaligned information .
Noising. The feature map extracted from an image occluded by obstacles and non-target pedestrians often contains some noise irrelevant to the target pedestrian, which is detrimental to the re-ID system for identification. To this end, we further perturb the feature map by noising to generate an adversarial representation with noisy information. Specifically, with the initialized , , and , we take a noising region and replace its pixels with random values that follow a uniform distribution. The value range of and , and the calculation way of and are the same as the above erasing process. Let denote the adversarial representation, it satisfies the following formula:
(4) |
Note that the erased, transformed, and noised regions are the same for all images in a mini-batch, which prevents the model from being hard to converge [19].
III-C Adversarial Defense
Existing adversarial defense studies on person re-ID mainly focus on training the model jointly with adversarial samples and original inputs, which is time-consuming and poorly defensive. Inspired by the adversarial training in GAN [21], which optimizes a generator and a discriminator in an adversarial way to encourage the former to generate fake samples that make the latter can not distinguish them from the real ones. For person re-ID, we expect the feature extractor to learn a sufficiently robust representation whose perturbed version can be correctly identified by the classifier as well. To this end, with the above adversarial representations, we develop three modules with a novel GAN-based adversarial defense strategy to overcome the issues of missing information, position misalignment, and noisy information caused by occlusion. No additional networks are required, each module employs adversarial training between the feature extractor and identity classifier, simple yet effective.
Erasing Defense. Compared to the clean feature map that can be correctly identified (Baseline Branch), the adversarial representation loses some identity-related cues due to the information of being erased, leading the classifier into misidentification. Therefore, we maximize the cross-entropy loss to optimize identity classifier :
(5) |
where is the probability that is recognized as category .
For the feature extractor , we expect the learned feature map from it can well depict the pedestrian characteristics, even if part of the information is lost, it can be correctly identified. To achieve this goal, we utilize to update the parameters of by minimizing the cross-entropy loss:
(6) |
Transforming Defense. Given the adversarial representation , since its information at is inconsistent with the information at on the feature map , the classifier fails to accurately identify the target pedestrian, which can be formulated as:
(7) |
where is the prediction result of belonging to the class .
Conversely, if the feature map enables the classifier to correctly identify its attacked version , it means that is not affected by the misaligned information. To this end, we constrain to satisfy the following formula:
(8) |
Noising Defense. In order to further endow with the ability to resist the interference of noisy information, similar to the above processes, we utilize the perturbed representation to maximize the cross-entropy loss for optimizing , and in turn, minimizing the cross-entropy loss for optimizing . Suppose that the probability of being correctly classified is , the above optimization strategy can be expressed as follows:
(9) |
(10) |
Note that in all the above adversarial learning processes, the feature extractor is fixed when training the identity classifier and vice versa. Through continuous gaming and optimization between them, the re-ID system is able to defend against the perturbation of various occlusion issues. To gain a better understanding of the above adversarial game process, we take the erasing defense module as an example to further illustrate. As shown in Figure 3, assuming that the erased area on the feature map is the part marked by the red box. Initially, we fix the feature extractor and optimize the classifier by maximizing the identity loss. This enhances the discriminative capability of the classifier’s class prototype and improves its correct classification threshold. Next, we fix the classifier and optimize the feature extractor by minimizing the identity loss, which ensures that even if the information is erased, the feature map extracted by the feature extractor can be correctly identified by the classifier. Achieving this requires ensuring that the model can focus on non-erased areas with highly discriminative information (such as the part marked by the blue box). Through continuous adversarial gaming between the two, we can mine as much discriminative information as possible, thus protecting the re-ID model from the disruption caused by missing information due to occlusion. Similarly, our transforming defense and noising defense modules can also resist the interference of position misalignment and noisy information.
III-D Training and Inference
During the training stage, we utilize the baseline learning module to extract discriminative features and three defense modules to enhance the occlusion robustness of the features. Correspondingly, the total loss can be formalized as:
(11) |
where , and and are similar. , , and are hyper-parameters utilized to control the relative importance of the three adversarial losses. The training process is carried out in an end-to-end manner, which is summarized in Algorithm 1.

In the test phase, we extract the feature map and process it with pooling and normalization. After that, the obtained feature vector is used for inference.
IV Experiments
In this section, we conduct extensive experiments to demonstrate the effectiveness and superiority of the proposed method. First, we introduce six public person re-ID datasets used in the experiments. And then, we elaborate on the implementation details and evaluation protocols. After that, we compare the recognition performance of our approach with the state-of-the-art methods. Next, we verify the effectiveness of each module and analyze the impact of the hyper-parameters on the recognition rate. Finally, we further evaluate the proposed method with a discussion part.
IV-A Datastes
Occluded-DukeMTMC [35] is a classical occluded person database derived from DukeMTMC-reID [52]. It contains 15,618 training images of 702 pedestrians and 19,871 testing images of 1,110 pedestrians. Due to various occlusions in both query and gallery images, this dataset is quite challenging.
Occluded-REID [53] contains 2,000 images of 200 identities, and each identity has five occluded images and five full-body images. Following [53, 19], we take occluded images as the queries and non-occluded images as the galleries, and randomly select half of them for training and the rest for testing.
P-DukeMTMC-reID [53] is another variant of DukeMTMC-reID that targets occluded person re-ID task. It consists of 12,927 training images of 665 identities and 11,216 testing images of 634 identities (2,163 occluded images in the query set and 9,053 holistic images in the gallery set).
Market-1501 [54] consists of 12,936 training images, 3,368 query images, and 19,732 gallery images. All images are captured from 6 non-overlapping cameras deployed at the campus. This dataset primarily targets the holistic person re-ID task due to few images being occluded.
DukeMTMC-reID is widely used to evaluate the performance of holistic person re-ID, which contains 36,411 images of 1,812 pedestrians collected from 8 cameras (408 distractor identities only appear in one camera view). All images are divided into 16,522 training samples and 19,889 test samples (This dataset is for research use only in this work).
MSMT17 [55] includes 126,441 images of 4,101 pedestrians, in which 32,621 images of 1,041 identities are used for training and the remaining are used for testing. This dataset is currently the most challenging holistic person database due to the long time span of sample collection.
More details about each dataset can be found in Table I.
Datasets | Train | Test(ID/Imgs) | |
---|---|---|---|
(ID/Imgs) | Query | Gallery | |
O-Duke | 702/15,618 | 519/2,210 | 1,110/17,661 |
Occluded-REID | 100/1,000 | 100/500 | 100/500 |
P-Duke | 665/12,927 | 634/2,163 | 634/9,053 |
Market-1501 | 751/12,936 | 750/3,368 | 750/19,732 |
DukeMTMC-reID | 702/16,522 | 702/2,228 | 1,110/17,661 |
MSMT17 | 1,041/32,621 | 3,060/11,659 | 3,060/82,161 |
Methods | Ref | Ext | Bac | Occluded-DukeMTMC | Occluded-REID | ||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Rank-1 | Rank-5 | Rank-10 | mAP | Rank-1 | Rank-5 | Rank-10 | mAP | ||||
PGFA [35] | ICCV’19 | ✓ | CNN | 51.4 | 68.6 | 74.9 | 37.3 | 57.1 | 77.9 | 84.0 | 56.2 |
PVPM [10] | CVPR’20 | ✓ | CNN | 47.0 | - | - | - | 70.4 | 84.1 | 89.8 | 64.2 |
HOReID [11] | CVPR’20 | ✓ | CNN | 55.1 | - | - | 43.8 | 80.3 | - | - | 70.2 |
GASM [56] | ECCV’20 | ✓ | CNN | - | - | - | - | 74.5 | - | - | 65.6 |
OAMN [57] | ICCV’21 | ✓ | CNN | 62.6 | 77.5 | - | 46.1 | - | - | - | - |
Pirt [58] | ACMM’21 | ✓ | CNN | 60.0 | - | - | 50.9 | - | - | - | - |
PGFL-KD [37] | ACMM’21 | ✓ | CNN | 63.0 | - | - | 54.1 | 80.7 | - | - | 70.3 |
PGMANet [59] | IJCNN’21 | ✓ | CNN | 51.3 | 66.5 | 73.4 | 40.9 | - | - | - | - |
RFCnet [12] | TPAMI’21 | ✓ | CNN | 63.9 | 77.6 | 82.1 | 54.5 | - | - | - | - |
PEFB [60] | TNNLS’21 | ✓ | CNN | 56.3 | 72.4 | 78.0 | 43.5 | - | - | - | - |
SORN [36] | TCSVT’21 | ✓ | CNN | 57.6 | 73.7 | 79.0 | 46.3 | - | - | - | - |
Adver Occluded [18] | CVPR’18 | ✗ | CNN | 44.5 | - | - | 32.2 | - | - | - | - |
PCB [27] | ECCV’18 | ✗ | CNN | 42.6 | 57.1 | 62.9 | 33.7 | 66.6 | 89.2 | - | - |
AFPB [53] | ICME’18 | ✗ | CNN | - | - | - | - | 68.1 | 88.3 | 93.7 | - |
FPR [34] | ICCV’19 | ✗ | CNN | - | - | - | - | 78.3 | - | - | 68.0 |
ISP [61] | ECCV’20 | ✗ | CNN | 62.8 | 78.1 | 82.9 | 52.3 | - | - | - | - |
RE [51] | AAAI’20 | ✗ | CNN | 40.5 | 59.6 | 66.8 | 30.0 | 65.8 | 87.9 | - | - |
RFCnet§ [12] | TPAMI’21 | ✗ | CNN | 52.4 | 68.3 | 73.4 | 44.8 | - | - | - | - |
IGOAS [19] | TIP’21 | ✗ | CNN | 60.1 | - | - | 49.4 | 81.1 | 91.6 | - | - |
ASAN [62] | TCSVT’21 | ✗ | CNN | 55.4 | 72.4 | 78.9 | 43.8 | 82.5 | 92.2 | - | 71.8 |
CBDBNet [63] | TCSVT’21 | ✗ | CNN | 50.9 | 66.0 | 74.2 | 38.9 | - | - | - | - |
FED [64] | CVPR’22 | ✗ | VIT | 68.1 | 79.3 | - | 56.4 | 86.3 | - | - | 79.3 |
QPM [65] | TMM’22 | ✗ | CNN | 64.4 | 79.3 | 84.2 | 49.7 | - | - | - | - |
DRL-Net [66] | TMM’22 | ✗ | VIT | 65.8 | 80.4 | 85.2 | 53.9 | - | - | - | - |
OPR-DAAO [67] | TIFS’22 | ✗ | CNN | 66.2 | 78.4 | 83.9 | 55.4 | 84.2 | 87.3 | - | 75.1 |
RTGTA [68] | TIP’23 | ✗ | CNN | 61.0 | 69.7 | 73.6 | 50.1 | 71.8 | 94.6 | 99.0 | 51.0 |
Baseline | - | ✗ | CNN | 52.5 | 69.3 | 75.0 | 44.6 | 87.8 | 95.3 | 97.5 | 79.5 |
Ours | - | ✗ | CNN | 63.6 | 79.5 | 84.7 | 54.7 | 89.3 | 96.0 | 97.9 | 81.5 |
Ours* | - | ✗ | CNN | 68.1 | 82.5 | 87.7 | 57.6 | 90.5 | 96.5 | 98.5 | 81.9 |
IV-B Settings
Implementation Details: We conduct all experiments on one GTX3090 with the Pytorch platform. All images are uniformly resized to 256128 and subjected to random flipping, padding, cropping, and erasing to achieve data augmentation. The batch size is set to 64, each batch contains 8 pedestrians and each pedestrian has 8 images. We set the initial learning rate to and decrease it by a factor of 0.1 at the 40th epoch and 70th epoch. The size of the feature map is , , and . The hyper-parameters are set to , , and . We utilize the Adam optimizer [69] to update the model parameters and train 120 epochs in total.
Methods | Bac | Rank-1 | Rank-5 | Rank-10 | mAP |
---|---|---|---|---|---|
PCB [27] | CNN | 79.4 | 87.1 | 90.0 | 63.9 |
IDE [70] | CNN | 82.9 | 89.4 | 91.5 | 65.9 |
PVPM [10] | CNN | 85.1 | 91.3 | 93.3 | 69.9 |
PGFA [35] | CNN | 85.7 | 92.0 | 94.2 | 72.4 |
ISP [61] | CNN | 89.0 | 94.1 | 95.3 | 74.7 |
Baseline | CNN | 91.2 | 95.0 | 95.9 | 76.5 |
Ours | CNN | 91.6 | 95.1 | 96.2 | 79.2 |
Ours* | CNN | 92.7 | 95.6 | 96.6 | 80.7 |
Evaluation Protocol: We follow the public evaluation indicators to verify the performance of the proposed method, namely Cumulative Matching Characteristics (CMC) and mean Average Precision (mAP). Note that all experiments are performed in the single query setting.
IV-C Comparison with State-of-the-art Methods
Experiments on occluded datasets: We first evaluate the superiority of the proposed method on the three public occluded datasets, namely Occluded-DukeMTMC, Occluded-REID, and P-DukeMTMC-reID. The comparison results are shown in Table II and Table III, where we report the accuracy of Rank-1, Rank-5, Rank-10, and mAP.
For Occluded-DukeMTMC, methods that rely on external models have achieved satisfactory results. For example, RFCNet [12] achieves 63.9% Rank-1 and 54.5% mAP by utilizing pose and semantic information. However, this superior accuracy is dependent on the reliability of external models and requires significant computational resources. As shown in the table, in the absence of auxiliary information, RFCNet only achieves 52.4% (-11.5%) Rank-1 and 44.8% (-9.7%) mAP. These results demonstrate the heavy reliance of such methods on external visual cues. In contrast, our method achieves comparable performance to RFCNet without utilizing external models and with minimal computational resources. Recently, model-free methods without external models have gained attention, and our proposed method falls into this category while having notable advantages in terms of performance and model simplicity. Specifically, the optimal ISP [61] predicts human body parts using clustering, avoiding the need for a pre-trained semantic parsing network, but it suffers from significant computational complexity and inferior performance compared to our method (52.3% vs. 54.7% mAP). Furthermore, Adver Occluded [18] and IGOAS [19], which utilize adversarial defense by generating aggressive samples, achieve only 44.5% and 60.1% Rank-1 accuracy, and 32.2% and 49.4% mAP, respectively, while our ETNDNet achieves 63.6% Rank-1 accuracy and 54.7% mAP. These results highlight the stronger occlusion defense ability of the model trained by our method. Additionally, the attention mechanism is commonly employed by state-of-the-art methods [64, 66] to extract feature maps containing crucial pedestrian information. In our proposed ETNDNet, adversarial representations formed by perturbing such feature maps are more aggressive, and leveraging them for adversarial defense enables the re-ID model to learn more robust features against occlusions. To achieve this, we incorporate Non-Local [71] into our backbone. As shown in Table II, our method (Ours*) achieves superior performance, surpassing the aforementioned methods (e.g., +1.2% mAP compared to FED [64]), further demonstrating the effectiveness and superiority of our approach.
For Occluded-REID, we repeat the experiment 10 times as in [19] and take the mean as the final result. As shown in Table II, the proposed method achieves satisfactory results. For example, compared with the optimal pose-guided based method PGFL-KD [37], the proposed ETNDNet improves the Rank-1 accuracy from 80.7% to 89.3%, and mAP from 70.3% to 81.5%. Compared with the optimal external model-free method FED [64], Rank-1 and mAP of ours are improved by 3.0% and 2.2%. Moreover, IGOAS [19], which is based on adversarial learning, the proposed ETNDNet surpasses it by 8.2% in Rank-1 accuracy. Finally, for P-DukeMTMC-reID, as shown in Table III, our method also outperforms state-of-the-art algorithms and achieves significant improvements in mAP, such as 6.8% over the optimal ISP [61].
According to the above results, the proposed method has significant advantages for occluded person re-ID, which is mainly attributed to the fact that our adversarial representations fully simulate perturbation caused by various occlusion issues (missing information, position misalignment, and noisy information), and the model trained by the developed GAN-based adversarial defense has the ability to protect the re-ID system from such disturbance, effectively overcoming the adverse effect of occlusion issues on recognition performance.
Experiments on holistic datasets: To further evaluate the superiority and practicality of our method, we compare with the state-of-the-art methods on holistic re-ID datasets, i.e., Market-1501, DukeMTMC-reID, and MSMT17, and report the comparison results in Table IV.
Methods | Bac | Market-1501 | DukeMTMC-reID | MSMT17 | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Rank-1 | Rank-5 | mAP | Rank-1 | Rank-5 | mAP | Rank-1 | Rank-5 | mAP | ||
PTGAN [55] | CNN | - | - | - | - | - | - | 68.2 | - | 40.4 |
PGFA [35] | CNN | 91.2 | - | 76.8 | 82.6 | - | 65.5 | - | - | - |
PEFB [60] | CNN | 92.7 | - | 81.3 | 86.2 | - | 72.6 | - | - | - |
VPM [72] | CNN | 93.0 | 97.8 | 80.8 | 83.6 | 91.7 | 72.6 | - | - | - |
IGOAS [19] | CNN | 93.4 | - | 84.1 | 86.9 | - | 75.1 | - | - | - |
HOReID [11] | CNN | 94.2 | - | 84.9 | 86.9 | - | 75.6 | - | - | - |
IANet [73] | CNN | 94.4 | - | 83.1 | 87.1 | - | 73.4 | 75.5 | 85.5 | 46.8 |
BagTricks [4] | CNN | 94.5 | - | 85.9 | 86.4 | - | 76.4 | - | - | - |
MHSANet [74] | CNN | 94.6 | - | 84.0 | 87.3 | - | 73.1 | - | - | - |
SORN [36] | CNN | 94.8 | - | 84.5 | 86.9 | - | 74.1 | - | - | - |
OSNet [75] | CNN | 94.8 | - | 84.9 | 88.6 | - | 73.5 | 78.7 | - | 52.9 |
AGW [5] | CNN | 95.1 | - | 87.8 | 89.0 | - | 79.6 | 78.3 | - | 55.6 |
POS [76] | CNN | 95.0 | 98.3 | 86.2 | 94.6 | 88.7 | 76.7 | - | - | - |
PLIP [77] | CNN | 95.1 | - | 88.0 | 86.5 | - | 77.0 | - | - | - |
BPBreID [78] | CNN | 95.1 | - | 87.0 | 89.6 | - | 78.3 | - | - | - |
GPS [79] | CNN | 95.2 | 98.4 | 87.8 | 88.2 | 95.2 | 78.7 | - | - | - |
Baseline | CNN | 93.8 | 98.1 | 84.1 | 86.1 | 93.2 | 74.2 | 74.8 | 85.2 | 49.0 |
Ours | CNN | 95.3 | 98.2 | 87.2 | 88.5 | 94.7 | 77.9 | 80.9 | 89.3 | 56.6 |
Ours* | CNN | 95.7 | 98.5 | 88.7 | 89.3 | 95.3 | 78.8 | 82.7 | 91.1 | 58.0 |
Although these three datasets are released for holistic person re-ID, due to the changeable postures and complex backgrounds, they still exist problems of information loss, information misalignment, and noisy information. Therefore, our method is also applicable and can further improve the recognition rate. Specifically, for Market-1501, our ETNDNet outperforms the state-of-the-art GPS [79] in terms of Rank-1 accuracy. Additionally, the integration of an attention mechanism into our framework led to a further improvement in recognition performance, with Rank-1 being 0.5% higher than GPS and mAP being 0.7% higher than PLIP [77]. As for DukeMTMC-reID, our ETNDNet also surpasses most state-of-the-art methods. Moreover, MSMT17 is currently the largest dataset, and it is relatively difficult due to the large difference between positive and negative pairs. Existing methods increase training samples (PTGAN [55]) or introduce attention mechanisms (IANet [73], OSNet [75], and AGW [5]) to reduce the intra-class distance. Even so, our method with less training cost achieves superior performance with 80.9% Rank-1 and 56.6% mAP, which is ahead of the state-of-the-art methods.
IV-D Model Analysis
Ablation Studies. We conduct ablation experiments to demonstrate the contribution of each module in the proposed ETNDNet, including erasing defense (ED), transforming defense (TD), and noising defense (ND). All experiments were performed on Occluded-DukeMTMC, which can better reflect the effectiveness of modules. The Rank-1 accuracy and mAP (%) are reported in Table V.
Effectiveness of ED. The developed ED is designed to protect the re-ID system from missing information. As shown in Table V, when deploying it to the baseline re-ID system, the Rank-1 accuracy and mAP are improved from 52.5% to 58.7% and from 44.6% to 49.7%, respectively. Moreover, compared with NDNet, the Rank-1 and mAP obtained by ENDNet have also been significantly improved. These results verify that our ED is indeed able to promote the re-ID model to learn features that are not disturbed by information loss, playing a positive role in identifying occluded target pedestrians and can cooperate with other modules to further improve the recognition performance.



Effectiveness of TD. The proposed TD strives to defend the re-ID model for extracting features unaffected by misaligned information. As shown in Table V, our TDNet promotes the Rank-1 accuracy and mAP of the Baseline by a large margin. And, when it is equipped on the above EDNet, the Rank-1 accuracy and mAP are improved from 58.7% to 62.1% and from 49.7% to 52.8%. These results demonstrate that our TD is also beneficial to occluded person re-ID, effectively alleviating the adverse impact of position misalignment caused by occlusion on recognition performance.
Methods | B | ED | TD | ND | Rank-1 | mAP |
---|---|---|---|---|---|---|
Baseline | ✓ | 52.5 | 44.6 | |||
EDNet | ✓ | ✓ | 58.7 | 49.7 | ||
TDNet | ✓ | ✓ | 57.5 | 49.7 | ||
NDNet | ✓ | ✓ | 58.0 | 48.8 | ||
ETDNet | ✓ | ✓ | ✓ | 62.1 | 52.8 | |
ENDNet | ✓ | ✓ | ✓ | 62.3 | 52.6 | |
TNDNet | ✓ | ✓ | ✓ | 59.0 | 51.2 | |
ETNDNet | ✓ | ✓ | ✓ | ✓ | 63.6 | 54.7 |
ETNDNet* | ✓ | ✓ | ✓ | ✓ | 68.1 | 57.6 |
Effectiveness of ND. The proposed ND is dedicated to solving the problem of occlusion noise introduced by obstacles and non-target pedestrians, which is ignored in existing studies. As shown in Table V, it also contributes to the improvement of recognition performance, and when it is combined with other modules, the formed framework can achieve higher accuracy. Specifically, our NDNet increases the Rank-1 accuracy of the re-ID baseline model by 5.5% and mAP by 4.2%. And, the performance of ENDNet and TNDNet is also higher than that of EDNet and TDNet. These results confirm our ND is indeed robust against the interference of noisy information.
Finally, when all modules are equipped, the developed ETNDNet can achieve the best recognition rates of 63.6% Rank-1 and 54.7% mAP. It follows that our three adversarial defense modules are able to collaborate with each other, effectively overcoming the negative effects of perturbation of various occlusion issues on identification.
Parameter Analysis. Hyper-parameters , , and are used to balance the relative importance of adversarial losses in the above three adversarial defense modules. Here, we study the influence of their different values on recognition performance. Each parameter was tested with 11 different values and the experimental results are illustrated in Figure 4. Note that when one of the parameters is analyzed, the remaining two are fixed at the optimal values.
The effect of . From Figure 4, we can see that when increases from 0 to 0.1, both Rank-1 accuracy and mAP are improved, and when , our method achieves the best recognition performance. This indicates that is an optimal value for the proposed method. Moreover, we observe that the recognition performance decreases as continues to increase, which may be due to over-guaranteeing the robustness of the learned features while destroying their discriminativeness.
The effect of . The hyper-parameter is used to control the relative importance of the adversarial loss in the proposed transforming defense module. As shown in Figure 4, the performance of the developed ETNDNet remains relatively stable when increases from 0 to 0.5. And, when , the recognition performance slightly exceeds the corresponding performance achieved at other values of . So, is set to 0.15 in all the experiments.
The effect of . In order to further improve the robustness of the learned features, the ND is designed and the parameter is used to control its importance. It can be seen that Rank-1 accuracy and mAP reach the peak when , which further confirms the validity of ND and the rationality of .
IV-E Further Discussions
In this section, we further evaluate the proposed method. All experiments were performed on Occluded-Duke with the same hardware (1 GTX3090 GPU, 8 CPU, and 32GB memory).
Erasing Strategy. In our erasing module, we perturb feature maps with a random erasing strategy to form adversarial representations with incomplete information. The strategy proposed in [80] can also serve this purpose. However, since it fixes the erased area’s size proportion and aspect ratio while different obstacles have various sizes and shapes in real scenes, perturbing feature maps by such strategy hinders the aggressiveness of adversarial representations and ultimately limits recognition performance improvement. As shown in Figure 5(a), when we follow DropBlock [80] to fix the erased area’s size proportion and aspect ratio to 0.3, the Rank-1 and mAP of our EDNet dropped by 1.3% and 3.0%. Obviously, it is more appropriate to adopt our random erasing strategy to perturb feature maps.


Defense Paradigm. In this paper, we design a novel GAN-based adversarial defense paradigm to protect the re-ID system from various occlusion disturbances, which is significantly different from the adversarial representation learning of IGOAS [19]. Specifically, IGOAS employs an adversarial suppression branch that implements a mean square loss between the feature map and its masked version to make the model focus on pedestrian information in non-erased areas. In contrast, our GAN-based adversarial defense paradigm continuously plays an adversarial game between the identity classifier and feature extractor, encouraging the model to focus on highly discriminative information in non-disturbed areas. Particularly, in the developed defense paradigm, optimizing the feature extractor alone without the adversarial game process is comparable to IGOAS’s learning approach (that is, treating adversarial representations as supervision signals to optimize the feature extractor, making the model focus on pedestrian information in non-disturbed areas). As demonstrated in Figure 5(b), removing the adversarial game process reduces the mAP and Rank-1 accuracy by 4.8% and 5.9%, respectively, indicating the superiority of the learning way in our GAN-based adversarial defense paradigm.
Model Complexity. To further demonstrate the practicality of the proposed method, we conduct an analysis of the model complexity and compare ours with four state-of-the-art methods. The results are illustrated in Table VI, where we report the parameter number of models (Param), the number of floating-point operations for an input image (FLOPs), and the inference time of retrieving all queries (IT). It can be observed that existing state-of-the-art methods have a relatively large model complexity. In comparison, our Param, FLOPs, and IT are only 24.9M, 4.1, and 25s, thus the proposed ETNDNet has a significant advantage in model simplicity. Moreover, our method does not introduce any external modules with learnable parameters. Therefore, it has the same model complexity as the baseline, that is to say, our method improves the recognition performance without increasing the computational burden, with strong practicability. Additionally, incorporating Non-Local into our backbone, ETNDNet* increases the number of parameters by only 0.2M, but leads to significant performance improvements, surpassing state-of-the-art methods.
Methods | Param | FLOPs | IT | Rank-1 | mAP |
HOReID [11] | 163.2M | 4.9 | 258s | 55.1 | 43.8 |
IGOAS [19] | 32.6M | 4.6 | 49s | 60.1 | 49.4 |
RFCnet [12] | 25.6M | 4.3 | 119s | 63.9 | 54.5 |
OPR-DAAO [67] | 58.2M | 9.2 | - | 66.2 | 55.4 |
FED [64] | 146.2M | 11.0 | 98s | 68.1 | 56.4 |
Ours(Baseline) | 24.9M | 4.1 | 25s | 52.5 | 44.6 |
Ours(ETNDNet) | 24.9M | 4.1 | 25s | 63.6 | 54.7 |
Ours(ETNDNet*) | 25.1M | 4.1 | 44s | 68.1 | 57.6 |
Model Generalization Ability. We also evaluate the model generalization ability of our algorithms, with the results presented in Table VII. Importantly, it should be noted that our ETNDNet was not primarily tailored for domain generalization, and its performance falls short of the state-of-the-art method [81]. Nonetheless, in comparison to other occluded person re-ID algorithms, our model demonstrates superior generalization ability. To illustrate, for Market Duke, our ETNDNet outperforms HOReID by 1.6% and IGOAS by 0.2% in terms of mAP. Furthermore, compared with the Baseline, our algorithm yields a substantial enhancement of 6.0% in Rank-1 accuracy and 4.1% in mAP. This observation underscores the efficacy of the proposed algorithm in enhancing model generalization.
Methods | Market Duke | Duke Market | ||
---|---|---|---|---|
Rank-1 | mAP | Rank-1 | mAP | |
QAConv [82] | 48.8 | 28.7 | 58.6 | 27.2 |
SNR [81] | 55.1 | 33.6 | 66.7 | 33.9 |
HOReID [11] | 39.4 | 23.2 | 48.0 | 22.1 |
IGOAS [19] | 41.5 | 24.6 | 48.7 | 22.2 |
Baseline | 35.1 | 20.7 | 56.2 | 26.9 |
Ours | 41.1 | 24.8 | 56.3 | 27.6 |
Ours* | 48.3 | 30.6 | 60.1 | 30.2 |
Adversarial Representation. Aggressiveness is an important basis for judging the quality of adversarial examples. As discussed earlier, our adversarial representations formed by perturbing feature maps are more aggressive than adversarial samples formed in [18, 19]. To demonstrate this, we test the performance of the model trained by the baseline framework on adversarial samples and various adversarial representations. The experimental results are shown in Figure 6 (a). Firstly, compared with the baseline, the adversarial samples reduce the mAP and Rank-1 by 9.3% and 8.0%, while our adversarial representations with incomplete information reduce the mAP and Rank-1 by 20.8% and 19.1%, which suggests that applying perturbation on the feature map is indeed more aggressive. Secondly, the adversarial representations generated by random transforming simulate the issue of position misalignment caused by occlusion, which is a capability that adversarial samples do not possess. And, as can be seen from Figure 6 (a), they lead to the reduction of the mAP and Rank-1 from 44.6% to 36.0% and from 52.5% to 46.1%. Thirdly, our adversarial representations with noisy information also reduce the recognition rate by a large margin (-16.3% mAP and -16.5% Rank-1) and are more effective in attacking the re-ID model than adversarial samples.


Model Defensive Capability. As we discussed above, the model trained by our ETNDNet has a stronger model defensive capability than IGOAS [19]. In order to further prove this, during testing, we randomly erase the query and gallery samples to further cheat the re-ID system and compare the recognition performance obtained by our method and IGOAS. As illustrated in Figure 6 (b), the performance that IGOAS achieved was far from the results of its baseline model. Different from it, the proposed ETNDNet gains a recognition rate comparable to the baseline. This is mainly because our adversarial defense strategy enables the re-ID model to pay close attention to as many pedestrian body parts as possible. Even if the information of a certain part is lost, the target pedestrian can still be correctly identified (as illustrated in Figure 7). The above qualitative and quantitative analysis fully demonstrate that our method with the developed novel GAN-based adversarial defense is valuable in object re-identification and deserves further exploration.






Limitations. In the proposed ETNDNet, we perturb feature maps by random erasing, transforming, and noising to form adversarial representations that can simulate various issues caused by occlusions. However, the erased, transformed, and noised areas possibly overlap with occlusions. Additionally, the shapes of obstacles and human bodies are often irregular, while our erased, transformed, and noised areas are regular rectangles, hindering the aggressiveness of adversarial representations and ultimately limiting performance improvement. These limitations prompt us to conduct further investigations in the future.
V Conclusion
In this paper, from the perspective of adversarial defense, we propose a simple yet effective framework named ETNDNet to protect the re-ID system from various occlusion perturbations. Specifically, we generate adversarial representations with incomplete information, misaligned information, and noisy information by erasing, transforming, and noising at feature maps to simulate issues of missing information, position misalignment, and noisy information caused by occlusion. Accordingly, we develop a novel GAN-based adversarial defense approach to train the feature extractor and identity classifier in an adversarial way, thus overcoming the detrimental effects of the aforementioned perturbations on the re-ID system. The proposed ETNDNet exhibits superior results on six public re-ID databases, and extensive experiments fully demonstrate its effectiveness and practicability. In the future, we will further explore the potential of adversarial attack and defense in object re-identification.
References
- [1] H. B. Shitrit, J. Berclaz, F. Fleuret, and P. Fua, “Multi-commodity network flow for tracking multiple people,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, no. 8, pp. 1614–1627, 2014, doi: 10.1109/TPAMI.2013.210.
- [2] Y. Li, L. Ma, Z. Zhong, F. Liu, M. A. Chapman, D. Cao, and J. Li, “Deep learning for lidar point clouds in autonomous driving: A review,” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 8, pp. 3412–3432, 2021, doi: 10.1109/TNNLS.2020.3015992.
- [3] J. Tang, X. Shu, R. Yan, and L. Zhang, “Coherence constrained graph lstm for group activity recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 2, pp. 636–647, 2022, doi: 10.1109/TPAMI.2019.2928540.
- [4] H. Luo, Y. Gu, X. Liao, S. Lai, and W. Jiang, “Bag of tricks and a strong baseline for deep person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2019, pp. 1487–1495, doi: 10.1109/CVPRW.2019.00190.
- [5] M. Ye, J. Shen, G. Lin, T. Xiang, L. Shao, and S. C. H. Hoi, “Deep learning for person re-identification: A survey and outlook,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 6, pp. 2872–2893, 2022, doi: 10.1109/TPAMI.2021.3054775.
- [6] X. Shu, X. Wang, X. Zang, S. Zhang, Y. Chen, G. Li, and Q. Tian, “Large-scale spatio-temporal person re-identification: Algorithms and benchmark,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 7, pp. 4390–4403, 2022, doi: 10.1109/TCSVT.2021.3128214.
- [7] H. Li, Y. Chen, D. Tao, Z. Yu, and G. Qi, “Attribute-aligned domain-invariant feature learning for unsupervised domain adaptation person re-identification,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 1480–1494, 2021, doi: 10.1109/TIFS.2020.3036800.
- [8] Y. Wang, G. Qi, S. Li, Y. Chai, and H. Li, “Body part-level domain alignment for domain-adaptive person re-identification with transformer framework,” IEEE Transactions on Information Forensics and Security, vol. 17, pp. 3321–3334, 2022, doi: 10.1109/TIFS.2022.3207893.
- [9] Y. Peng, S. Hou, C. Cao, X. Liu, Y. Huang, and Z. He, “Deep learning-based occluded person re-identification: A survey,” arXiv preprint arXiv:2207.14452, 2022.
- [10] S. Gao, J. Wang, H. Lu, and Z. Liu, “Pose-guided visible part matching for occluded person reid,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 11 744–11 752, doi: 10.1109/CVPR42600.2020.01176.
- [11] G. Wang, S. Yang, H. Liu, Z. Wang, Y. Yang, S. Wang, G. Yu, E. Zhou, and J. Sun, “High-order information matters: Learning relation and topology for occluded person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 6449–6458, doi: 10.1109/CVPR42600.2020.00648.
- [12] R. Hou, B. Ma, H. Chang, X. Gu, S. Shan, and X. Chen, “Feature completion for occluded person re-identification,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 9, pp. 4894–4912, 2022, doi: 10.1109/TPAMI.2021.3079910.
- [13] Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh, “Realtime multi-person 2d pose estimation using part affinity fields,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 7291–7299, doi: 10.1109/CVPR.2017.143.
- [14] X. Liang, K. Gong, X. Shen, and L. Lin, “Look into person: Joint body parsing & pose estimation network and a new benchmark,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 4, pp. 871–885, 2018, doi: 10.1109/TPAMI.2018.2820063.
- [15] I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” arXiv preprint arXiv:1412.6572, 2014.
- [16] S. Zheng, Y. Song, T. Leung, and I. Goodfellow, “Improving the robustness of deep neural networks via stability training,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 4480–4488, doi: 10.1109/CVPR.2016.485.
- [17] C. Lyu, K. Huang, and H.-N. Liang, “A unified gradient regularization family for adversarial examples,” in Proceedings of the IEEE International Conference on Data Mining, 2015, pp. 301–309, doi: 10.1109/ICDM.2015.84.
- [18] H. Huang, D. Li, Z. Zhang, X. Chen, and K. Huang, “Adversarially occluded samples for person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 5098–5107, doi: 10.1109/CVPR.2018.00535.
- [19] C. Zhao, X. Lv, S. Dou, S. Zhang, J. Wu, and L. Wang, “Incremental generative occlusion adversarial suppression network for person reid,” IEEE Transactions on Image Processing, vol. 30, pp. 4212–4224, 2021, doi: 10.1109/TIP.2021.3070182.
- [20] A. Ilyas, S. Santurkar, D. Tsipras, L. Engstrom, B. Tran, and A. Madry, “Adversarial examples are not bugs, they are features,” Advances in Neural Information Processing Systems, vol. 32, 2019.
- [21] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” Communications of the ACM, vol. 63, no. 11, pp. 139–144, 2020, doi: 10.1145/3422622.
- [22] H. Lee, S. Han, and J. Lee, “Generative adversarial trainer: Defense to adversarial perturbations with gan,” arXiv preprint arXiv:1705.03387, 2017.
- [23] M. Farenzena, L. Bazzani, A. Perina, V. Murino, and M. Cristani, “Person re-identification by symmetry-driven accumulation of local features,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2010, pp. 2360–2367, doi: 10.1109/CVPR.2010.5539926.
- [24] S. Liao, Y. Hu, X. Zhu, and S. Z. Li, “Person re-identification by local maximal occurrence representation and metric learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 2197–2206, doi: 10.1109/CVPR.2015.7298832.
- [25] H. Li, S. Yan, Z. Yu, and D. Tao, “Attribute-identity embedding and self-supervised learning for scalable person re-identification,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 30, no. 10, pp. 3472–3485, 2019, doi: 10.1109/TCSVT.2019.2952550.
- [26] D. Yi, Z. Lei, S. Liao, and S. Z. Li, “Deep metric learning for person re-identification,” in Proceedings of the IEEE International Conference on Pattern Recognition (ICPR), 2014, pp. 34–39, doi: 10.1109/ICPR.2014.16.
- [27] Y. Sun, L. Zheng, Y. Yang, Q. Tian, and S. Wang, “Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline),” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 480–496, doi: 10.1007/978-3-030-01225-0_30.
- [28] Z. Jin, J. Xie, B. Wu, and L. Shen, “Weakly supervised pedestrian segmentation for person re-identification,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 33, no. 3, pp. 1349–1362, 2023, doi: 10.1109/TCSVT.2022.3210476.
- [29] S. Li, F. Li, K. Wang, G. Qi, and H. Li, “Mutual prediction learning and mixed viewpoints for unsupervised-domain adaptation person re-identification on blockchain,” Simulation Modelling Practice and Theory, vol. 119, p. 102568, 2022, doi: 10.1016/j.simpat.2022.102568.
- [30] Y. Wang, K. Xu, Y. Chai, Y. Jiang, and G. Qi, “Semantic consistent feature construction and multi-granularity feature learning for visible-infrared person re-identification,” The Visual Computer, pp. 1–17, 2023, doi: 10.1007/s00371-023-02923-w.
- [31] H. Li, N. Dong, Z. Yu, D. Tao, and G. Qi, “Triple adversarial learning and multi-view imaginative reasoning for unsupervised domain adaptation person re-identification,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 5, pp. 2814–2830, 2022, doi: 10.1109/TCSVT.2021.3099943.
- [32] J. Zhuo, J. Lai, and P. Chen, “A novel teacher-student learning framework for occluded person re-identification,” arXiv preprint arXiv:1907.03253, 2019.
- [33] M. Kiran, R. G. Praveen, L. T. Nguyen-Meidine, S. Belharbi, L.-A. Blais-Morin, and E. Granger, “Holistic guidance for occluded person re-identification,” arXiv preprint arXiv:2104.06524, 2021.
- [34] L. He, Y. Wang, W. Liu, H. Zhao, Z. Sun, and J. Feng, “Foreground-aware pyramid reconstruction for alignment-free occluded person re-identification,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2019, pp. 8450–8459, doi: 10.1109/ICCV.2019.00854.
- [35] J. Miao, Y. Wu, P. Liu, Y. Ding, and Y. Yang, “Pose-guided feature alignment for occluded person re-identification,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2019, pp. 542–551, doi: 10.1109/ICCV.2019.00063.
- [36] X. Zhang, Y. Yan, J.-H. Xue, Y. Hua, and H. Wang, “Semantic-aware occlusion-robust network for occluded person re-identification,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 7, pp. 2764–2778, 2020, doi: 10.1109/TCSVT.2020.3033165.
- [37] K. Zheng, C. Lan, W. Zeng, J. Liu, Z. Zhang, and Z.-J. Zha, “Pose-guided feature learning with knowledge distillation for occluded person re-identification,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 4537–4545, doi: 10.1145/3474085.3475610.
- [38] H. Tang, Z. Li, Z. Peng, and J. Tang, “Blockmix: meta regularization and self-calibrated inference for metric-based meta-learning,” in Proceedings of the 28th ACM International Conference on Multimedia, 2020, pp. 610–618, doi: 10.1145/3394171.3413884.
- [39] H. Tang, C. Yuan, Z. Li, and J. Tang, “Learning attention-guided pyramidal features for few-shot fine-grained recognition,” Pattern Recognition, vol. 130, p. 108792, 2022, doi: 10.1016/j.patcog.2022.108792.
- [40] Z. Li, H. Tang, Z. Peng, G.-J. Qi, and J. Tang, “Knowledge-guided semantic transfer network for few-shot image recognition,” IEEE Transactions on Neural Networks and Learning Systems, 2023, doi: 10.1109/TNNLS.2023.3240195.
- [41] Z. Zha, H. Tang, Y. Sun, and J. Tang, “Boosting few-shot fine-grained recognition with background suppression and foreground alignment,” IEEE Transactions on Circuits and Systems for Video Technology, 2023, doi: 10.1109/TCSVT.2023.3236636.
- [42] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus, “Intriguing properties of neural networks,” arXiv preprint arXiv:1312.6199, 2013.
- [43] A. Kurakin, I. Goodfellow, and S. Bengio, “Adversarial machine learning at scale,” arXiv preprint arXiv:1611.01236, 2016.
- [44] Z. Wang, S. Zheng, M. Song, Q. Wang, A. Rahimpour, and H. Qi, “advpattern: physical-world attacks on deep person re-identification via adversarially transformable patterns,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2019, pp. 8341–8350, doi: 10.1109/ICCV.2019.00843.
- [45] H. Wang, G. Wang, Y. Li, D. Zhang, and L. Lin, “Transferable, controllable, and inconspicuous adversarial attacks on person re-identification with deep mis-ranking,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 342–351, doi: 10.1109/CVPR42600.2020.00042.
- [46] J. Gao, B. Wang, Z. Lin, W. Xu, and Y. Qi, “Deepcloak: Masking deep neural network models for robustness against adversarial samples,” arXiv preprint arXiv:1702.06763, 2017.
- [47] X. Wang, S. Li, M. Liu, Y. Wang, and A. K. Roy-Chowdhury, “Multi-expert adversarial attack detection in person re-identification using context inconsistency,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2021, pp. 15 097–15 107, doi: 10.1109/ICCV48922.2021.01482.
- [48] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778, doi: 10.1109/CVPR.2016.90.
- [49] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009, pp. 248–255, doi: 10.1109/CVPR.2009.5206848.
- [50] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 2818–2826, doi: 10.1109/CVPR.2016.308.
- [51] Z. Zhong, L. Zheng, G. Kang, S. Li, and Y. Yang, “Random erasing data augmentation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 07, 2020, pp. 13 001–13 008, doi: 10.1609/aaai.v34i07.7000.
- [52] E. Ristani, F. Solera, R. Zou, R. Cucchiara, and C. Tomasi, “Performance measures and a data set for multi-target, multi-camera tracking,” in Proceedings of the European Conference on Computer Vision (ECCV). Springer, 2016, pp. 17–35, doi: 10.1007/978-3-319-48881-3_2.
- [53] J. Zhuo, Z. Chen, J. Lai, and G. Wang, “Occluded person re-identification,” in IEEE International Conference on Multimedia and Expo (ICME), 2018, pp. 1–6, doi: 10.1109/ICME.2018.8486568.
- [54] L. Zheng, L. Shen, L. Tian, S. Wang, J. Wang, and Q. Tian, “Scalable person re-identification: A benchmark,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2015, pp. 1116–1124, doi: 10.1109/ICCV.2015.133.
- [55] L. Wei, S. Zhang, W. Gao, and Q. Tian, “Person transfer gan to bridge domain gap for person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 79–88, doi: 10.1109/CVPR.2018.00016.
- [56] L. He and W. Liu, “Guided saliency feature learning for person re-identification in crowded scenes,” in Proceedings of the European Conference on Computer Vision (ECCV). Springer, 2020, pp. 357–373, doi: 10.1007/978-3-030-58604-1_22.
- [57] P. Chen, W. Liu, P. Dai, J. Liu, Q. Ye, M. Xu, Q. Chen, and R. Ji, “Occlude them all: Occlusion-aware attention network for occluded person re-id,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 11 813–11 822, doi: 10.1109/ICCV48922.2021.01162.
- [58] Z. Ma, Y. Zhao, and J. Li, “Pose-guided inter-and intra-part relational transformer for occluded person re-identification,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1487–1496, doi: 10.1145/3474085.3475283.
- [59] Y. Zhai, X. Han, W. Ma, X. Gou, and G. Xiao, “Pgmanet: Pose-guided mixed attention network for occluded person re-identification,” in 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, 2021, pp. 1–8, doi: 10.1109/IJCNN52387.2021.9534442.
- [60] J. Miao, Y. Wu, and Y. Yang, “Identifying visible parts via pose estimation for occluded person re-identification,” IEEE Transactions on Neural Networks and Learning Systems, 2021, doi: 10.1109/TNNLS.2021.3059515.
- [61] K. Zhu, H. Guo, Z. Liu, M. Tang, and J. Wang, “Identity-guided human semantic parsing for person re-identification,” in Proceedings of the European Conference on Computer Vision (ECCV). Springer, 2020, pp. 346–363, doi: 10.1007/978-3-030-58580-8_21.
- [62] H. Jin, S. Lai, and X. Qian, “Occlusion-sensitive person re-identification via attribute-based shift attention,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 4, pp. 2170–2185, 2021, doi: 10.1109/TCSVT.2021.3088446.
- [63] H. Tan, X. Liu, Y. Bian, H. Wang, and B. Yin, “Incomplete descriptor mining with elastic loss for person re-identification,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 1, pp. 160–171, 2021, doi: 10.1109/TCSVT.2021.3061412.
- [64] Z. Wang, F. Zhu, S. Tang, R. Zhao, L. He, and J. Song, “Feature erasing and diffusion network for occluded person re-identification,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4754–4763, doi: 10.1109/CVPR52688.2022.00471.
- [65] P. Wang, C. Ding, Z. Shao, Z. Hong, S. Zhang, and D. Tao, “Quality-aware part models for occluded person re-identification,” IEEE Transactions on Multimedia, 2022, doi: 10.1109/TMM.2022.3156282.
- [66] M. Jia, X. Cheng, S. Lu, and J. Zhang, “Learning disentangled representation implicitly via transformer for occluded person re-identification,” IEEE Transactions on Multimedia, 2022, doi: 10.1109/TMM.2022.3141267.
- [67] S. Wang, R. Liu, H. Li, G. Qi, and Z. Yu, “Occluded person re-identification via defending against attacks from obstacles,” IEEE Transactions on Information Forensics and Security, vol. 18, pp. 147–161, 2022, doi: 10.1109/TIFS.2022.3218449.
- [68] M. Huang, C. Hou, Q. Yang, and Z. Wang, “Reasoning and tuning: Graph attention network for occluded person re-identification,” IEEE Transactions on Image Processing, vol. 32, pp. 1568–1582, 2023, doi: 10.1109/TIP.2023.3247159.
- [69] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- [70] L. Zheng, H. Zhang, S. Sun, M. Chandraker, Y. Yang, and Q. Tian, “Person re-identification in the wild,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 1367–1376, doi: 10.1109/CVPR.2017.357.
- [71] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7794–7803, doi: 10.1109/CVPR.2018.00813.
- [72] Y. Sun, Q. Xu, Y. Li, C. Zhang, Y. Li, S. Wang, and J. Sun, “Perceive where to focus: Learning visibility-aware part-level features for partial person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 393–402, doi: 10.1109/CVPR.2019.00048.
- [73] R. Hou, B. Ma, H. Chang, X. Gu, S. Shan, and X. Chen, “Interaction-and-aggregation network for person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 9317–9326, doi: 10.1109/CVPR.2019.00954.
- [74] H. Tan, X. Liu, B. Yin, and X. Li, “Mhsa-net: Multihead self-attention network for occluded person re-identification,” IEEE Transactions on Neural Networks and Learning Systems, 2022, doi: 10.1109/TNNLS.2022.3144163.
- [75] K. Zhou, Y. Yang, A. Cavallaro, and T. Xiang, “Omni-scale feature learning for person re-identification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 3702–3712, doi: 10.1109/ICCV.2019.00380.
- [76] S. Wang, B. Huang, H. Li, G. Qi, D. Tao, and Z. Yu, “Key point-aware occlusion suppression and semantic alignment for occluded person re-identification,” Information Sciences, vol. 606, pp. 669–687, 2022, doi: 10.1016/j.ins.2022.05.077.
- [77] J. Zuo, C. Yu, N. Sang, and C. Gao, “Plip: Language-image pre-training for person representation learning,” arXiv preprint arXiv:2305.08386, 2023.
- [78] V. Somers, C. De Vleeschouwer, and A. Alahi, “Body part-based representation learning for occluded person re-identification,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 1613–1623, doi: 10.1109/WACV56688.2023.00166.
- [79] B. X. Nguyen, B. D. Nguyen, T. Do, E. Tjiputra, Q. D. Tran, and A. Nguyen, “Graph-based person signature for person re-identifications,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 3492–3501, doi: 10.1109/CVPRW53098.2021.00388.
- [80] Z. Dai, M. Chen, X. Gu, S. Zhu, and P. Tan, “Batch dropblock network for person re-identification and beyond,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 3691–3701, doi: 10.1109/ICCV.2019.00379.
- [81] X. Jin, C. Lan, W. Zeng, Z. Chen, and L. Zhang, “Style normalization and restitution for generalizable person re-identification,” in proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 3143–3152, doi: 10.1109/CVPR42600.2020.00321.
- [82] S. Liao and L. Shao, “Interpretable and generalizable person re-identification with query-adaptive convolution and temporal lifting,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XI 16. Springer, 2020, pp. 456–474, doi: 10.1007/978-3-030-58621-8_27.