POISE: Pose Guided Human Silhouette Extraction under Occlusions
Abstract
Human silhouette extraction is a fundamental task in computer vision with applications in various downstream tasks. However, occlusions pose a significant challenge, leading to incomplete and distorted silhouettes. To address this challenge, we introduce POISE : Pose Guided Human Silhouette Extraction under Occlusions, a novel self-supervised fusion framework that enhances accuracy and robustness in human silhouette prediction. By combining initial silhouette estimates from a segmentation model with human joint predictions from a 2D pose estimation model, POISE leverages the complementary strengths of both approaches, effectively integrating precise body shape information and spatial information to tackle occlusions. Furthermore, the self-supervised nature of POISE eliminates the need for costly annotations, making it scalable and practical. Extensive experimental results demonstrate its superiority in improving silhouette extraction under occlusions, with promising results in downstream tasks such as gait recognition. The code for our method is available https://github.com/take2rohit/poise.
1 Introduction
Human silhouette extraction is a fundamental task in computer vision with wide-ranging applications in human motion analysis [30], surveillance [36], augmented reality [1], and human-computer interaction [2]. The precise extraction of human silhouettes from images and videos enables the accomplishment of higher-level tasks, including gait recognition [9] and activity recognition [18]. However, the presence of occlusions [3] poses substantial challenges to the extraction process, resulting in incomplete or distorted silhouettes that impede the performance of downstream tasks.

Occlusions are common in real-world scenarios, where human subjects navigate complex and cluttered environments. They occur when certain areas of the human body are concealed or overlapped by objects or other individuals in the scene. Examples include objects obstructing body parts and self-occlusions caused by body parts coming into contact. State-of-the-art techniques for extracting human silhouettes, which involve pretrained semantic segmentation models like DeepLabv3 [5, 6], struggle in such scenarios. Despite being trained on large-scale natural datasets, these models are not explicitly trained to handle occlusions. Consequently, they tend to misidentify occluded regions as background, resulting in fragmented and inaccurate silhouette predictions, as shown in Figure 1. Additionally, it is infeasible to acquire datasets with complete ground-truth human segmentation masks under occlusions due to visual uncertainty.
To overcome these challenges, we present a novel fusion framework, Pose Guided Silhouette Extraction under Occlusions (POISE ), that significantly enhances the accuracy and robustness of human silhouette prediction. Our approach combines the predictions from a segmentation model, which initially estimates the silhouettes, with the human joint predictions from a 2D human pose estimation model. The pose estimation model provides crucial spatial information by predicting the keypoints and inferring the body structure, aiding in handling occlusions. However, it lacks detailed information about the body shape. This is addressed via the initial silhouettes predicted by the segmentation model for the unoccluded regions, capturing precise body shape information. By fusing these two complementary sources of information, we refine the silhouette predictions, resulting in silhouettes that are unfragmented and preserve body shape, as illustrated in Figure 1.
Existing approaches for jointly learning poses and silhouettes primarily rely on supervised learning [28, 50] which necessitates large annotated datasets for training. In this work, we relax this assumption and instead learn the fusion framework in an self-supervised fashion. Specifically, given pretrained segmentation and pose estimation models, we train the fusion model to mimic the predictions given by the pair via a pseudo-labeling approach. In order to utilize the pose predictions for silhouette estimation, we design a novel auxiliary transformation function to convert the sparse keypoint estimates into dense human silhouettes. Our self-supervised approach alleviates the need for costly annotations and makes the method more scalable and applicable to real-world scenarios.
Main contributions. Our primary contributions are summarized as follows:
-
1.
We address the problem of self-supervised human silhouette extraction under occlusion.
-
2.
We present an effective fusion framework that leverages pretrained 2D pose estimation and silhouette extraction models to produce accurate and robust human silhouettes.
-
3.
Our framework is based on self-supervised learning, which eliminates the need for costly pixel-level annotations or pose annotations, thereby enhancing scalability.
-
4.
Beyond showcasing excellent results on silhouette extraction benchmarks, we also demonstrate the utility of the predicted silhouettes on downstream tasks such as gait recognition.
The rest of the paper is organized as follows: in Section 2, we review prior works in human silhouette extraction, pose estimation, and gait recognition. Section 3 describes the proposed framework in detail. Experimental results and analysis are presented in Section 4, followed by conclusions and future directions in Section 5.
2 Related Works
Human Parsing and Silhouette Extraction. Extensive research has been conducted on human silhouette extraction, with earlier works such as [52] utilizing Hidden Markov models for modeling human silhouettes. More recent approaches leverage deep learning by primarily addressing the problem as semantic segmentation [6, 26, 5, 51, 29, 34, 13, 4]. These methods have achieved high accuracy, but require large labeled datasets of human images with corresponding segmentation masks. Pretrained models like DeepLabV3 [6] perform well in diverse settings due to their training on extensive datasets. Despite the success of these methods, occlusion remains a challenge. Zhou et. al. [55] proposed a multistage architecture for de-occluding humans, while Li et. al. [22] introduced a state-of-the-art human parsing framework that can also be utilized for silhouette extraction by mapping all parts of the human body to the foreground. However, these algorithms require labeled training data, which is a limiting factor.
Human Pose Estimation. Human pose estimation involves localizing keypoints on the human body, such as the head, elbows and knees, in 2D or 3D space. Deep learning-based pose estimation methods such as [44, 37, 27, 46], have achieved remarkable success on challenging academic datasets. However, these models are typically trained in supervised settings and often exhibit limited generalization capabilities when applied to unseen images. To overcome this, Zhang et. al. [48] proposed a novel domain adaptive 3D pose estimation algorithm. Jiang et. al. [16] proposed RegDA, a domain adaptive 2D pose estimation algorithm, which was further improved upon by Kim et. al. [17] in their work on UDAPE. Further, a source-free approach was recently proposed in [33]. These unsupervised methods have made significant contributions to enhancing pose estimation in settings where labeled data is unavailable.
Human Pose Estimation under Occlusions: Modern human pose estimation algorithms often fail to localize human keypoints under occlusions. To counter the same, Zhou et. al. [54] introduced a novel algorithm based on siamese networks and feature matching to improve 2D human pose estimation performance under occlusions. Cheng et. al. [8] exploits spatio-temporal continuity for handing occlusions thus leading to improved pose estimation. Qiu et. al. [31] used a novel graph formulation for improving human pose estimation under multi-person occlusion scenario. Liu et. al. [25] introduced a novel multi-stage framework for obtaining human keypoints under occlusions. Note that, these algorithms are entirely supervised learning algorithms necessitating costly annotations. To mitigate the same, Wang et. al. [41] introduced a novel contrastive learning based occlusion-aware algorithm for predicting 3D human pose from given 2D keypoints. However, this algorithm still needs access to ground-truth 2D keypoints.
Multi-Task Learning for Pose and Parsing. Multitask learning for human pose and body parsing aims to leverage the complementary information from both tasks to enhance their individual performances. Nie et. al. [28] introduced an adaptive convolutional architecture that facilitates joint human pose estimation and body parsing. Their approach was trained in a supervised manner, minimizing a linear combination of losses for each task. Liang et. al. [23] proposed a novel architecture that utilizes the semantic correlation between pose estimation and body parsing tasks. By exploiting this correlation, they were able to improve the accuracy of both tasks. More recently, Zhang et. al. [50] explored the explicit cross-task consistency between pose estimation and parsing. However, all of these methods are fully supervised and thus, require labeled data for both tasks.
Gait Recognition. Gait recognition, a fundamental task in computer vision, involves identifying individuals based on their unique walking patterns. Silhouettes have been extensively studied and utilized for gait recognition for nearly two decades. Pioneering works by Collins et. al. [9] and Wang et. al. [42] laid the groundwork for employing silhouettes in this context. Since then, numerous studies [40] have explored the use of silhouettes for gait recognition. As gait recognition under occlusions poses a challenge, it is crucial to preprocess silhouettes to handle occlusions before incorporating them into gait recognition systems. This preprocessing step ensures the robustness and reliability of the gait recognition process, even in the presence of occlusions.

3 Method
3.1 Problem Formulation
Given a dataset of images , where each , our goal is to train a model that accurately predicts the silhouette corresponding to an input image . To achieve this, we leverage the predictions of pretrained segmentation () and 2D pose estimation () models. These models provide initial predictions for the silhouette and sparse keypoints, respectively.
In order to align the sparse keypoints from with dense human silhouettes, we utilize a pose-to-silhouette transformation function , where K represents the number of keypoints. We term this function Pose2Sil. By carefully fusing the predictions from and the transformed keypoints using , POISE aims to produce accurate and robust silhouettes.
3.2 Pose Guided Silhouette Extraction
For each image , we generate two silhouettes, and , using pretrained models and , respectively. Specifically,
(1a) | ||||
(1b) |
Here, represents the silhouette obtained from the pre-trained segmentation model, while is derived from the sparse pose predictions transformed by .
These pair of silhouettes provide complementary information. While retains image-specific body shape details, it may become fragmented in the presence of occlusions. On the other hand, produces a continuous silhouette but may lose accuracy in capturing the body shape. Thus, our problem of obtaining robust silhouettes under occlusions boils down to learning a mixed representation of the two individually noisy silhouettes. To learn a robust silhouette representation, we train to effectively combine the information from these two noisy silhouettes under occlusions. This is accomplished by minimizing the pixel-wise binary cross-entropy loss between the predicted silhouette and those obtained from the pretrained models,
(2a) | ||||
(2b) |
Notably, we train our model to focus exclusively on the foreground information from and ignore the background. This selective attention allows to concentrate solely on the identity-specific features derived from . This is particularly crucial since the segmentation model may misidentify body parts as background due to occlusion, rather than the other way around. In practice, this is carried out via a simple masking operation, which ignores all background pixels in .
We also use the pseudo-labels obtained from itself to regularize the training. Given an image , we first obtain the pseudo-label . These pseudo-labels are subsequently used to train the model using the pixel-wise binary cross-entropy loss,
(3) |
In order to prevent noisy predictions from impeding the learning process, we use a confidence threshold to mask out possible incorrect pseudo-labels. This ensures only the high-quality predictions are reinforced by the model.
The overall training objective is given by
(4) |
where each , for , controls the influence of individual loss terms in generating the final silhouette. An overview of our framework can be found in Figure 2.

3.3 Obtaining Silhouette from Pose Keypoints
To leverage the valuable information derived from the pose keypoints for training , we incorporate a dedicated module, Pose2Sil, which facilitates the transformation from pose to silhouette. In this section, we outline the training methodology employed for this module.
We assume access to an auxiliary synthetic dataset containing images of humans with corresponding annotations for pose keypoints and segmentation masks . Due to its synthetic nature, it is simple to obtain annotations [39], unlike the real-world images in . We train by minimizing the binary cross-entropy loss between the silhouette predicted from the pose and the corresponding ground-truth silhouette, . We illustrate this training process in Figure 3. Since the model does not rely on RGB images during training or inference, it remains unaffected by domain changes. Thus, it is domain-agnostic and can be applied seamlessly across different domains without retraining.
3.4 Estimating Pose Keypoints under Occlusion
Due to the limited generalization capabilities of pretrained pose estimation models [16, 17], directly applying them to the images in may result in subpar keypoint localization, as illustrated in figure 4. Consequently, self-supervised domain adaptation of these pretrained models becomes imperative to ensure their effectiveness on our specific domain.
In this work, we build upon the state-of-the-art domain adaptive pose estimation algorithm, UDAPE [17]. We leverage the availability of a synthetic dataset as our source dataset, and serves as our target dataset. The images in may contain occlusions, which can significantly hinder the accuracy of pose estimation, as highlighted in previous research [54]. To enhance pose estimation under occlusions on the target dataset, we introduce occlusions into the source dataset itself [20]. By doing so, we force the model to learn robust representations that facilitate improved pose estimation under occlusions in the target dataset. Please note that we do not have access to the ground-truth labels of .

The adaptation algorithm follows the self-training framework inspired by the widely-used Mean-Teacher approach [38]. It involves the utilization of two identical models: a teacher model and a student model. Both models are initially initialized with the same weights at time step . Subsequently, at each time step , the student model parameters are updated by leveraging the supervisory signals provided by the teacher model, as well as the annotated data from the synthetic dataset . The parameters of the teacher model, denoted as , are updated using an exponential moving average (EMA) of the student model parameters, ensuring a smoother and more stable learning process.

To update the student model, a combination of supervised and self-supervised losses is employed. The supervised loss is computed using a mean square criterion on the source dataset, leveraging the annotated ground-truth labels. In addition to the supervised loss, a self-supervised consistency criterion is introduced to encourage consistency in the pose predictions of two different augmentations of an image. Figure 5 shows a schematic diagram of the adaptation algorithm for learning . Once the adaptation is done, we use the updated teacher model as to obtain the pose estimates on the images in . Figure 6 shows the importance of introducing occlusions in the source data which leads to improved pose estimation under occlusions.

4 Experiments and Results
In this section, we provide a thorough assessment of POISE , highlighting its exceptional ability to accurately extract human silhouettes even when they are partially obscured. We evaluate POISE on five datasets, assessing its performance not only in human silhouette extraction but also in gait recognition. Our method outperforms existing off-the-shelf solutions and requires no extra annotations.
4.1 Datasets
We use the following datasets in our experiments.
-
•
Humans3.6M [15] is large scale real-world video dataset with over 3 million frames featuring from 11 professional actors performing different actions such as walking, eating, etc.. Following standard protocol, we use subjects ’S1’, ’S5’, ’S6’, ’S7’ and ’S8’ for training and subjects ’S9’ and ’S11’ for testing. Similar to [17], we use 20,000 frames for training and another 3000 for evaluation. We use the mean-IoU (mIoU) metric [35] to report segmentation performance on this dataset.
-
•
UP-S31 [21] is an image-based human part segmentation dataset with over 8000 images with 31 corresponding part annotations. However, several images containing more than one person were removed from our experiments leading to a total of 5226 images being used for our experiments. This was dis-jointly split into 4227 images for training and 1059 images for testing. We use the mean-IoU (mIoU) metric [35] to report segmentation performance on this dataset.
-
•
CASIA-B [45] is a large-scale multi-view indoor gait recognition dataset with 124 subjects across 11 views. Each subject has a total 10 sequences - six sequences under Normal (NM) conditions and two each under Carrying Bag (BG) and Wearing different Clothing (CL) conditions. We adhere to existing works such as [43] for training/testing and gallery/probe partitions. We report Rank-1 accuracy for gait recognition experiments using GaitBase [11] on this dataset.
-
•
BRIAR [10] is a recent large-scale real-world biometric dataset with images of individuals under challenging conditions such as atmospheric turbulence and natural occlusions. For this work, we prepare the frames by tracking the person with Byte-Track [49]. We then select about 20,000 frames at different distances, allocating roughly 75% for training and the remainder for testing. Since we lack ground-truth silhouettes, we present qualitative silhouette extraction results. Additionally, we evaluate gait recognition using GaitBase [11] on a subset of 87 subjects divided into 60 for training and 27 for testing.
-
•
3DOH50K [47] is large-scale real-world occlusion dataset with over 50000 frames. As the ground-truth segmentation masks are incomplete and images do not have ids associated with them, we report qualitative results for silhouette extraction on this dataset.
Generating occluded images. We use the datasets mentioned above to generate corresponding occluded RGB images for evaluation purposes. In particular, we consider two kinds of occlusions - Random Erase (RE) Occlusions [53] and Common Objects in Context (COCO) Occlusions [24]. For both Random Erase and COCO Occlusions, we first estimate keypoints on clean images using the adapted pose estimation model described in Section 3.4 and then add occlusions on randomly selected keypoints. This ensures that the occlusion always covers a certain part of the human body. Additional details on the same is presented in the section 1 of the supplementary.
4.2 Implementation details
In our experiments, we employ the DeepLabv3 architecture [7] with the ResNet-101 [14] feature extractor as the backbone for our model . For most of our experiments, we leverage the ResNet-101 pre-trained on the COCO dataset [24] as the segmentation network . However, for the specific scenario of COCO occlusions on Huamns3.6M and UP-S31 dataset, we adopt the SCHP architecture [22] pre-trained on the LIP dataset [12] as . For experiments with Random Erase Occlusions, is trained with five different severities of occlusion (12, 16, 20, 24, and 28), and the same model is used for inference on the five different severities of occlusion. Data augmentation strategies such as random rotation, translation, and shear are also used to regularize the training process. Additional implementation details are provided in the section 2 of supplementary.
Synthetic Dataset. We use SURREAL [39] as our synthetic dataset . SURREAL is a large-scale dataset containing over 6 million frames of synthetically generated human images against an indoor background.
Pose Estimation Model (). Adhering to [17], we use the Simple Baseline decoder [44] with Resnet-101 backbone [14] the architectures for our student and teacher networks. The networks are trained for a total of 80 epochs with the first 40 epochs being used for supervised training of the student model and the other 40 epochs for adaptation to the target domain. We use a batch size 32 while optimizing using the Adam optimizer [19] with an initial learning rate of , decaying by a factor of after and epochs.
Pose2Sil Model (). The architecture of is the same as that of the DCGAN [32]. The model is trained for a total of 200 epochs using the Adam optimizer [19] with a learning rate of and a batch size of 32.
Method | mIoU |
---|---|
80.14 | |
75.97 | |
POISE | 87.22 |
POISE + Weak Sup. | 89.67 |
Full Sup. | 92.75 |
Occlusion Severity | POISE | ||
---|---|---|---|
RE - 12 | 79.67 | 81.98 | 85.58 |
RE - 16 | 78.25 | 81.72 | 85.35 |
RE - 20 | 76.71 | 81.43 | 85.11 |
RE - 24 | 74.88 | 81.34 | 84.55 |
RE - 28 | 73.28 | 80.84 | 83.66 |
COCO | 78.68 | 77.34 | 80.86 |
4.3 Results
4.3.1 Silhouette Extraction
We evaluate the effectiveness of POISE for human silhouette extraction under occlusion in terms of segmentation accuracy. Table 1 shows the efficacy of POISE is learning strong feature representations to perform human silhouette extraction under COCO occlusions on Humans3.6M dataset. While our results are inferior to a fully-supervised baseline by 5.5 %, we show that we can bridge this gap to 3 % by fine-tuning the network with limited supervision, i.e. considering 5 % of the training dataset as annotated. This shows that POISE learns generalized feature representations that can be effectively used to obtain optimal human silhouettes under occlusions. We report the mIoU on UP-S31 dataset at six different occlusion severities in Table 2. We note that for Random Erase occlusions, POISE provides an improvement of against and against . For COCO occlusions, POISE provides an improvement of against and against . This shows the deficiencies of using state-of-the-art human segmentation methods in scenarios involving occlusion. We present additional qualitative results on the same in the section 3 of the supplementary.
We also assess the performance of POISE on the BRIAR dataset under natural occlusion scenarios. In Figure 7, we visually demonstrate how POISE excels at extracting silhouettes, closely resembling the original body shape from . In occluded regions, POISE relies on for guidance. Further, as shown in Figure 7, when dealing with human subjects of diverse body shapes (compared to the synthetic dataset ), fails to retain body-specific information, while retains most of it. POISE combines information from both and to produce optimal silhouettes.
Figure 8 shows the efficacy of POISE in handling natural occlusions on the 3DOH50K dataset. As POISE uses complementary self-supervisory signals (from and ) during training, it is able to handle occlusions reasonably well. The results are particularly interesting as it shows silhouettes from an self-supervised method i.e. POISE is better off as compared against ground-truth silhouettes. Additional results using recent segmentation methods are provided in section 4 of the supplementary.


Method | 20% | 33% | 50% | 67% | 80% | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NM | BG | CL | NM | BG | CL | NM | BG | CL | NM | BG | CL | NM | BG | CL | |
84.44 | 65.58 | 48.68 | 82.35 | 64.42 | 47.51 | 79.75 | 61.19 | 44.58 | 75.45 | 57.38 | 40.99 | 71.34 | 54.31 | 39.97 | |
54.95 | 33.22 | 25.75 | 52.08 | 31.64 | 24.47 | 51.72 | 30.68 | 23.36 | 49.44 | 29.88 | 24.40 | 48.44 | 29.66 | 23.50 | |
POISE | 83.38 | 68.38 | 52.63 | 83.01 | 68.15 | 52.49 | 81.85 | 66.87 | 51.01 | 81.05 | 65.15 | 49.82 | 78.96 | 63.54 | 49.17 |
Method | 12 | 16 | 20 | 24 | 28 | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NM | BG | CL | NM | BG | CL | NM | BG | CL | NM | BG | CL | NM | BG | CL | |
79.75 | 61.19 | 44.58 | 79.57 | 61.99 | 45.03 | 80.26 | 61.03 | 43.32 | 79.54 | 59.99 | 42.29 | 79.62 | 59.15 | 41.96 | |
51.72 | 30.68 | 23.36 | 48.78 | 30.21 | 22.56 | 46.75 | 28.31 | 22.03 | 44.72 | 27.12 | 21.46 | 43.23 | 25.93 | 20.75 | |
POISE | 81.85 | 66.87 | 51.01 | 81.25 | 65.52 | 50.08 | 80.47 | 66.32 | 51.11 | 80.05 | 64.70 | 49.57 | 80.36 | 65.73 | 49.02 |
Method | top-1 | top-3 | top-5 |
---|---|---|---|
8.96 | 22.39 | 31.72 | |
14.20 | 28.60 | 38.68 | |
POISE | 16.66 | 31.69 | 40.54 |
Method | NM | BG | CL | Avg. | |||
---|---|---|---|---|---|---|---|
✓ | 80.22 | 62.26 | 43.31 | 61.93 | |||
✓ | 52.23 | 31.07 | 24.79 | 36.03 | |||
POISE | ✓ | ✓ | 81.33 | 63.89 | 54.35 | 66.52 | |
POISE (full) | ✓ | ✓ | ✓ | 83.30 | 67.00 | 52.11 | 67.47 |
4.3.2 Gait Recognition
In this section, we present the efficacy of POISE on the downstream task of gait recognition on two datasets: CASIA-B and BRIAR.
Results on CASIA-B: We present gait recognition results on the CASIA-B dataset for three different walking conditions: Normal (NM), Carrying Bag (BG) and Clothing (CL). We evaluate POISE under the following two different settings.
Duration of Occlusion. In this setting the severity of the occlusion remains the same across frames but the number of occluded frames changes. As an example, occlusion means that for every video in the dataset, a temporally continuous set of frames of length of the total number of frames are occluded. Table 3 presents quantitative results on the performance of POISE against and at fixed occlusion severity of 12. It is interesting to note that POISE outperforms both and by significant margins under scenarios of temporally long occlusions.
Severity of Occlusion. Similar to our experimental settings for section 4.3.1, we study the effect of varying occlusion severity on gait recognition, keeping the same duration of occlusion. Table 4 presents quantitative results on the performance of POISE against and at an occlusion duration of . We note that, although with increasing severity of occlusion gait recognition performance falls for all the three listed methods, POISE still outperforms and by significant margins.
Results on BRIAR: Table 5 shows the efficacy of POISE in obtaining robust silhouettes that help with gait recognition on the BRIAR dataset. We obtain improvements of 7 % and 8 % in terms of top-1 and top-5 gait recognition accuracy over simply using . These improvements are significant as the videos in the dataset suffer from atmospheric turbulence and have instances of persistent natural occlusions.
A thorough discussion with additional quantitative analysis on gait recognition on the two aforementioned datasets is provided in section 5 of the supplementary.
4.3.3 Importance of Pseudo-Labeling:
5 Conclusion
We present POISE , an innovative approach for robust human silhouette extraction under occlusions. By leveraging pose estimation and self-supervised learning, POISE effectively addresses the limitations posed by occlusions, ensuring accurate and preserved body shape representations. Unlike traditional methods, POISE eliminates the need for expensive annotations, making it cost-effective and practical. Experimental results demonstrate the superiority of POISE in improving silhouette extraction under occlusions. Furthermore, POISE exhibits promising performance in gait recognition tasks, underscoring its potential impact in various applications. In summary, POISE offers a straightforward yet valuable solution for extracting precise silhouettes in challenging scenarios, contributing to advancements in biometrics and related fields.
Acknowledgements. This research is partially supported by the Office of the Director of National Intelligence (ODNI), specifically through the Intelligence Advanced Research Projects Activity (IARPA), under contract number [2022-21102100007]. The views and conclusions in this research reflect those of the authors and should not be construed as officially representing the policies, whether explicitly or implicitly, of ODNI, IARPA, or the U.S. Government. Nevertheless, the U.S. Government retains the authorization to reproduce and distribute reprints for official government purposes, regardless of any copyright notices included. We further thank our colleague Yash Garg ([email protected]) for helping us with additional experiments.
References
- [1] Yinoussa Adagolodjo, Raffaella Trivisonne, Nazim Haouchine, Stéphane Cotin, and Hadrien Courtecuisse. Silhouette-based pose estimation for deformable organs application to surgical augmented reality. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 539–544. IEEE, 2017.
- [2] Biplab Ketan Chakraborty, Debajit Sarma, Manas Kamal Bhuyan, and Karl F MacDorman. Review of constraints on vision-based gesture recognition for human–computer interaction. IET Computer Vision, 12(1):3–15, 2018.
- [3] Changhong Chen, Jimin Liang, Heng Zhao, Haihong Hu, and Jie Tian. Frame difference energy image for gait recognition with incomplete silhouettes. Pattern Recognition Letters, 30(11):977–984, 2009.
- [4] Hao Chen, Kunyang Sun, Zhi Tian, Chunhua Shen, Yongming Huang, and Youliang Yan. Blendmask: Top-down meets bottom-up for instance segmentation, 2020.
- [5] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L. Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs, 2017.
- [6] Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam. Rethinking atrous convolution for semantic image segmentation, 2017.
- [7] Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European conference on computer vision (ECCV), pages 801–818, 2018.
- [8] Yu Cheng, Bo Yang, Bo Wang, and Robby T Tan. 3d human pose estimation using spatio-temporal networks with explicit occlusion training. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 10631–10638, 2020.
- [9] Robert T Collins, Ralph Gross, and Jianbo Shi. Silhouette-based human identification from body shape and gait. In Proceedings of fifth IEEE international conference on automatic face gesture recognition, pages 366–371. IEEE, 2002.
- [10] David Cornett, Joel Brogan, Nell Barber, Deniz Aykac, Seth Baird, Nicholas Burchfield, Carl Dukes, Andrew Duncan, Regina Ferrell, Jim Goddard, et al. Expanding accurate person recognition to new altitudes and ranges: The briar dataset. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 593–602, 2023.
- [11] Chao Fan, Junhao Liang, Chuanfu Shen, Saihui Hou, Yongzhen Huang, and Shiqi Yu. Opengait: Revisiting gait recognition towards better practicality. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9707–9716, 2023.
- [12] Ke Gong, Xiaodan Liang, Dongyu Zhang, Xiaohui Shen, and Liang Lin. Look into person: Self-supervised structure-sensitive learning and a new benchmark for human parsing. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 932–940, 2017.
- [13] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn, 2018.
- [14] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [15] Catalin Ionescu, Dragos Papava, Vlad Olaru, and Cristian Sminchisescu. Human3.6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(7):1325–1339, jul 2014.
- [16] Junguang Jiang, Yifei Ji, Ximei Wang, Yufeng Liu, Jianmin Wang, and Mingsheng Long. Regressive domain adaptation for unsupervised keypoint detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6780–6789, 2021.
- [17] Donghyun Kim, Kaihong Wang, Kate Saenko, Margrit Betke, and Stan Sclaroff. A unified framework for domain adaptive pose estimation. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXXIII, pages 603–620. Springer, 2022.
- [18] Kibum Kim, Ahmad Jalal, and Maria Mahmood. Vision-based human activity recognition system using depth silhouettes: A smart home system for monitoring the residents. Journal of Electrical Engineering & Technology, 14:2567–2573, 2019.
- [19] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- [20] Jogendra Nath Kundu, Siddharth Seth, Pradyumna YM, Varun Jampani, Anirban Chakraborty, and R Venkatesh Babu. Uncertainty-aware adaptation for self-supervised 3d human pose estimation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 20448–20459, 2022.
- [21] Christoph Lassner, Javier Romero, Martin Kiefel, Federica Bogo, Michael J. Black, and Peter V. Gehler. Unite the people: Closing the loop between 3d and 2d human representations. In IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), July 2017.
- [22] Peike Li, Yunqiu Xu, Yunchao Wei, and Yi Yang. Self-correction for human parsing. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(6):3260–3271, 2020.
- [23] Xiaodan Liang, Ke Gong, Xiaohui Shen, and Liang Lin. Look into person: Joint body parsing & pose estimation network and a new benchmark. IEEE transactions on pattern analysis and machine intelligence, 41(4):871–885, 2018.
- [24] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740–755. Springer, 2014.
- [25] Qihao Liu, Yi Zhang, Song Bai, and Alan Yuille. Explicit occlusion reasoning for multi-person 3d human pose estimation. In European Conference on Computer Vision, pages 497–517. Springer, 2022.
- [26] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3431–3440, 2015.
- [27] Julieta Martinez, Rayat Hossain, Javier Romero, and James J Little. A simple yet effective baseline for 3d human pose estimation. In Proceedings of the IEEE international conference on computer vision, pages 2640–2649, 2017.
- [28] Xuecheng Nie, Jiashi Feng, and Shuicheng Yan. Mutual learning to adapt for joint human parsing and pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 502–517, 2018.
- [29] Hyeonwoo Noh, Seunghoon Hong, and Bohyung Han. Learning deconvolution network for semantic segmentation, 2015.
- [30] Ronald Poppe. Vision-based human motion analysis: An overview. Computer vision and image understanding, 108(1-2):4–18, 2007.
- [31] Lingteng Qiu, Xuanye Zhang, Yanran Li, Guanbin Li, Xiaojun Wu, Zixiang Xiong, Xiaoguang Han, and Shuguang Cui. Peeking into occluded joints: A novel framework for crowd pose estimation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIX 16, pages 488–504. Springer, 2020.
- [32] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
- [33] Dripta S Raychaudhuri, Calvin-Khang Ta, Arindam Dutta, Rohit Lal, and Amit K Roy-Chowdhury. Prior-guided source-free domain adaptation for human pose estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 14996–15006, 2023.
- [34] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks, 2016.
- [35] Hamid Rezatofighi, Nathan Tsoi, JunYoung Gwak, Amir Sadeghian, Ian Reid, and Silvio Savarese. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 658–666, 2019.
- [36] Vinay Sharma and James W Davis. Extraction of person silhouettes from surveillance imagery using mrfs. In 2007 IEEE Workshop on Applications of Computer Vision (WACV’07), pages 33–33. IEEE, 2007.
- [37] Ke Sun, Bin Xiao, Dong Liu, and Jingdong Wang. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5693–5703, 2019.
- [38] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Advances in neural information processing systems, 30, 2017.
- [39] Gul Varol, Javier Romero, Xavier Martin, Naureen Mahmood, Michael J Black, Ivan Laptev, and Cordelia Schmid. Learning from synthetic humans. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 109–117, 2017.
- [40] Changsheng Wan, Li Wang, and Vir V Phoha. A survey on gait recognition. ACM Computing Surveys (CSUR), 51(5):1–35, 2018.
- [41] Junjie Wang, Zhenbo Yu, Zhengyan Tong, Hang Wang, Jinxian Liu, Wenjun Zhang, and Xiaoyan Wu. Ocr-pose: Occlusion-aware contrastive representation for unsupervised 3d human pose estimation. In Proceedings of the 30th ACM International Conference on Multimedia, pages 5477–5485, 2022.
- [42] Liang Wang, Tieniu Tan, Huazhong Ning, and Weiming Hu. Silhouette analysis-based gait recognition for human identification. IEEE transactions on pattern analysis and machine intelligence, 25(12):1505–1518, 2003.
- [43] Zifeng Wu, Yongzhen Huang, Liang Wang, Xiaogang Wang, and Tieniu Tan. A comprehensive study on cross-view gait based human identification with deep cnns. IEEE transactions on pattern analysis and machine intelligence, 39(2):209–226, 2016.
- [44] Bin Xiao, Haiping Wu, and Yichen Wei. Simple baselines for human pose estimation and tracking. In Proceedings of the European conference on computer vision (ECCV), pages 466–481, 2018.
- [45] Shiqi Yu, Daoliang Tan, and Tieniu Tan. A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In 18th international conference on pattern recognition (ICPR’06), volume 4, pages 441–444. IEEE, 2006.
- [46] Feng Zhang, Xiatian Zhu, and Mao Ye. Fast human pose estimation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3517–3526, 2019.
- [47] Tianshu Zhang, Buzhen Huang, and Yangang Wang. Object-occluded human shape and pose estimation from a single color image. In IEEE Conference on Computer Vision and Pattern Recognition, (CVPR), 2020.
- [48] Xiheng Zhang, Yongkang Wong, Mohan S Kankanhalli, and Weidong Geng. Unsupervised domain adaptation for 3d human pose estimation. In Proceedings of the 27th ACM International Conference on Multimedia, pages 926–934, 2019.
- [49] Yifu Zhang, Peize Sun, Yi Jiang, Dongdong Yu, Fucheng Weng, Zehuan Yuan, Ping Luo, Wenyu Liu, and Xinggang Wang. Bytetrack: Multi-object tracking by associating every detection box. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXII, pages 1–21. Springer, 2022.
- [50] Ziwei Zhang, Chi Su, Liang Zheng, and Xiaodong Xie. Correlating edge, pose with parsing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8900–8909, 2020.
- [51] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network, 2017.
- [52] San-Lung Zhao and Hsi-Jian Lee. Human silhouette extraction based on hmm. In 18th International Conference on Pattern Recognition (ICPR’06), volume 2, pages 994–997, 2006.
- [53] Zhun Zhong, Liang Zheng, Guoliang Kang, Shaozi Li, and Yi Yang. Random erasing data augmentation. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 13001–13008, 2020.
- [54] Lu Zhou, Yingying Chen, Yunze Gao, Jinqiao Wang, and Hanqing Lu. Occlusion-aware siamese network for human pose estimation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XX 16, pages 396–412. Springer, 2020.
- [55] Qiang Zhou, Shiyin Wang, Yitong Wang, Zilong Huang, and Xinggang Wang. Human de-occlusion: Invisible perception and recovery for humans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3691–3701, 2021.