This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Weakly-supervised Pre-training for 3D Human Pose Estimation via Perspective Knowledge

Zhongwei Qiu Kai Qiu Jianlong Fu Dongmei Fu
Abstract

Modern deep learning-based 3D pose estimation approaches require plenty of 3D pose annotations. However, existing 3D datasets lack diversity, which limits the performance of current methods and their generalization ability. Although existing methods utilize 2D pose annotations to help 3D pose estimation, they mainly focus on extracting 2D structural constraints from 2D poses, ignoring the 3D information hidden in the images. In this paper, we propose a novel method to extract weak 3D information directly from 2D images without 3D pose supervision. Firstly, we utilize 2D pose annotations and perspective prior knowledge to generate the relationship of that keypoint is closer or farther from the camera, called relative depth. We collect a 2D pose dataset (MCPC) and generate relative depth labels. Based on MCPC, we propose a weakly-supervised pre-training (WSP) strategy to distinguish the depth relationship between two points in an image. WSP enables the learning of the relative depth of two keypoints on lots of in-the-wild images, which is more capable of predicting depth and generalization ability for 3D human pose estimation. After fine-tuning on 3D pose datasets, WSP achieves state-of-the-art results on two widely-used benchmarks.

keywords:
Human Pose Estimation, Pre-training, Relative Depth, Weakly-supervised.

1 Introduction

The goal of 3D human pose estimation is to estimate the 3D coordinates of human joints from 2D image, which provides accurate and fine-grained posture of the human body in the real 3D world. As an essential task in computer vision, 3D human pose estimation can be applied in human behavior understanding [1, 2], human-object interaction detection [3], and athletic training assistance [4] and so on. Recently, with the breakthrough of deep convolutional neural network (CNN) [5, 6] and the disclosure of some large-scale datasets [7, 8, 9] for human pose estimation, many 3D human pose estimation methods [10, 11, 12, 13, 14] have achieved remarkable improvements. However, the challenges of lacking rich 3D pose data and improving generalization ability still remain.

Monocular 3D human pose estimation heavily relies on large amounts of images and relative annotations of 3D human pose. At present, most 3D datasets of human pose are collected by a multi-view motion capture system (MoCap), which requires fine adjustment of equipment parameters and actors performing action performances in constraint places. Thus, collecting and labeling 3D human pose data consumes a lot of costs. Meanwhile, due to the limitation of the MoCap system, existing 3D pose estimation datasets [7, 8, 9] are collected in indoor lab or outdoor restricted space with several actors, which lack the diversity of background and human poses. A well-known issue [15] is that insufficient diversity of the training datasets will seriously affect the performance of 3D pose estimation network. The models trained on these constraint images usually fail to generalize in the real wild scenes, which limits the application of 3D human pose estimation models in the industry.

To solve the above problems, many works [16, 10, 13, 11, 14] train 3D pose models with the helping of some 2D pose datasets (e.g. MPII [17], COCO [18]). These mixing training strategy (MTS) can be summarized as Figure 1 (a). They use 2D heatmaps as the supervision of the network when inputting images without 3D annotation, while 3D pose as the supervision when inputting images with 3D annotation. Benefited from over ten thousand images provided by these 2D datasets [17, 18], annotated with accurate 2D pose in the natural scenes, they make the breakthrough of 3D human pose estimation. However, they just use the 2D pose as the supervision when training 3D pose models since no 3D annotations in these 2D datasets, ignoring the 3D information hidden in 2D images. For example, as shown in Figure 1 (b), human can distinguish the relations of person PAP_{A} is closer from the camera than person PBP_{B}. This weak 3D information can not be captured by just giving the supervision of 2D pose from single person. Other works have explored to alleviate the above problems by composting more training images [19, 20, 21, 8] or weakly supervised methods [13, 22, 23] by using multi-view and projecting information. However, the composited images remain gap in realism with natural images, resulting in different data distribution between training data and testing data. Besides, camera parameters and multi-view images can not be provided in existing 2D datasets, which limits the application of these weakly supervised methods.

Refer to caption
Figure 1: (a) Existing mixing training strategy (MTS) [10, 11] train network by mixing images with and without 3D pose annotations. (b) Our weakly-supervised pre-training (WSP) approach train network on images without 3D pose annotation in a weakly-supervised way, with the supervision of relative depth. Then, we finetune network on images with 3D pose annotations. (c) The process of generating relative depths via perspective effect.

In this paper, we tackle the two problems: 1) the problems of lacking enough 3D human pose annotation, 2) improving the generalization ability of 3D pose estimation model. Based on the observations of existing 2D human pose datasets, these real images can provide rich information of different background, diverse human poses, and different scales of human instances, which could help 3D human pose estimation. Moreover, we found that human could extract some weak 3D information directly from monocular 2D images by perspective knowledge. According to perspective effect, for an image captured by camera at a horizontal angle on the ground, the objects closer from camera are bigger than the objects farther from the camera. Meanwhile, for the points on the ground in this image, the points closer from camera have bigger YY-axis values in image coordinate system, which is named as relative depth.

As shown in Figure 1 (c), given a 2D image with multiple persons, it’s easy for human to distinguish which person is closer from the camera. They make the verdict based on the empirical knowledge of foreshortening effects. Therefore, we explore to solve the above problems by utilizing large scale 2D human pose data and perspective knowledge. Additional 3D information and rich posture data in natural scenes, provided from 2D datasets, can improve the accuracy of 3D pose estimation and the generalization ability of 3D pose estimation model. Different from previous methods [16, 10, 24, 13, 11, 25], we extract weak 3D information directly from 2D images without 3D annotations or other extra manual labeling, which can significantly improve the performance of predicting 3D pose without extra costs of data collection and labeling.

In order to learn relative depth of keypoints from 2D images, we collect a lager scale 2D dataset of human pose (MCPC) and generate relative depth according to the perspective knowledges. Since existing 2D dataset (MPII [17], COCO [18], PoseTrack [26] and CrowdPose [27]) don’t meet the requirements that multiple persons are in an image and each person has the annotations of ankle, we rearranged the images from the 2D datasets and labeled the annotations of relative depths of human bodies for each image by the perspective prior knowledge. Then, we propose a weakly-supervised pre-training (WSP) approach to learn weak 3D representation from MCPC dataset. WSP include two steps: pre-training on MCPC dataset and finetuning on 3D pose datasets. For pre-training, WSP firstly learns 2D heatmaps, which is used as spatial mask to extract keypoints feature. Then, WSP learns the relative depth of two keypoints from the keypoints features. For 3D human pose estimation, we finetune WSP on the 3D human pose estimation dataset. Since WSP learned weak 3D representation from a lot of real 2D images in the wild scenes, it can greatly improve the accuracy of 3D human pose estimation, especially in the depth prediction of 3D pose. In addition, WSP has strong generalization ability because it’s pre-trained on a lot of real 2D images with diverse human poses and background in the wild scenes.

Our main contributions can be summarized as follows:

  • 1.

    To the best of our knowledge, we are the first one, who uses the perspective knowledge to generate relative depths of keypoints for the pre-training of 3D human pose estimation. It’s less costly since no need of 3D pose annotation or artificial relative depth labeling.

  • 2.

    We propose a large-scale dataset (MCPC), with over 53k images and 220k instances, with the annotations of relative depth. These data could help relieve the need for large amounts of 3D pose annotations and improve the generalization ability of the 3D pose model.

  • 3.

    We propose the WSP approach and pre-train it on MCPC dataset. After finetuning WSP on Human3.6M and MuCo-3DHP datasets, our approach outperforms previous approaches and achieve state-of-the-art results on these two benchmarks.

The remainder of this paper is organized as follows. We introduce the related work in Section II. Our approach and MCPC dataset is introduced in Section III. Experiments and analysis are presented in Section IV. The Conclusions are summarized in Section V.

2 Related Work

We study the problems of 3D human pose estimation and try to improve the generalization ability of 3D pose estimation model. Here, we briefly review the related works on these problems.

2.1 3D human pose estimation

The progress in the field of 3D pose estimation in recent years benefit a lot from the development of deep learning. These methods mainly fall into two paradigms, directly inferring from images or lifting from 2D poses.

The lifting paradigm first conducts 2D human pose estimation on 2D images and then lifts 2D pose to 3D pose by a shallow network. Since there are a lot mature methods of 2D human pose estimation [28, 29, 6, 30, 31, 32, 33, 34], many lifting works [35, 36, 37, 38, 15] are mainly focus the lifting stage. Martinez et al. [35] proposes a simple fully connected network to lift the 3D pose from the 2D pose. Based on this work, some works study the influence of human pose structure by designing local or global network structures. Ci et al. [39] proposes a locally connected network to learn the local joint relations. Moreover, pose structure can be represented in a graph structure, with structural feature learning and 3D human pose estimation by graph convolutional network (GCN). Zhao et al. [38] proposes semantic GCN to capture the semantic information of human pose. These methods are lightweight and can be benefited from state-of-the-art 2D human pose models. However, the results of the lifting methods awfully depend on the quality of the 2D pose. The learned network is over-fitting on datasets without seeing the real 2D images. Pavlakos et al. [24] proposed to use a weaker supervision signal provided by the ordinal depths of human joints in the first stage. The estimated depths 2D coordinates are combined together as the input of lifinglifing network.

The inferring methods directly regress 3D coordinates of human joints from the input 2D image. Sun et al. [10] proposes to learn 3D heatmaps by a backbone network (ResNet), then regressing 3D coordinates by integral regression on 3D heatmaps, which is the simplest and state-of-the-art approach. This is the mainstream framework of inferring-based 3D human pose estimation. Based on this framework, many works try to study self-supervised methods [13], multi-view methods [12, 14] and multi-person 3D human pose estimation [11, 40]. Compared with lifting methods, the inferring methods study extracting features from the 2D images in an end-to-end way without overfitting on 2D pose data in a special dataset, which has greater application potential.

2.2 The problem of lacking 3D pose annotation

The success of existing 3D human pose estimation methods relies on multiple 3D datasets [8, 7, 9] with a larger number of 3D pose annotations. However, most of the existing 3D datasets are collected in the indoor lab or restricted outdoor space with several actors, who perform specific action performances. Therefore, the performance of existing 3D human pose estimation methods is limited by insufficient data on the 3D pose. The mainstream ways to address this problem are adding 2D pose datasets, generating more training images, and weakly supervised learning.

2.2.1 Mixing 2D pose dataset training

Some works [16, 10, 41, 42] have studied the mixed training of 2D and 3D human pose datasets. Zhou et al. [16] a weakly-supervised transfer learning network with a 2D pose estimation sub-network and a 3D depth regression sub-network, which uses mixed 2D and 3D labels. Sun et al. [10] proposed a unified framework to mix 2D and 3D labels by integral regression on 3D heatmaps. Yang et al. [42] proposed an adversarial learning framework, which distills the 3D human pose structures learned from the fully annotated dataset to in-the-wild images with only 2D pose annotations. Habibie et al. [41] introduced a network architecture that comprises a new disentangled hidden space encoding of explicit 2D and 3D features and uses supervision by a new learned projection model from a predicted 3D pose. These methods have extensively verified that 2D labels can improve the accuracy of 3D pose estimation. However, these methods just extract 2D information from images and 2D labels to enhance the constraint of the 3D pose, which limits the role of 2D labels and images. In this paper, We directly extract 3D information from 2D labels and images to improve the performance of 3D human pose estimation.

2.2.2 Data synthesis and data augmentation

Some approaches try to generate more training images through data augmentation [21, 43, 19, 8]. Varol et al. [21] propose to generate more training images by computer graphics rendering. Rogez et al. [43] propose mocap-guided data augmentation for 3D human pose estimation in the wild. Pishchulin et al. [19] fitted 3D models to humans and deformed them to generate new images in multiple views. Mehta et al. [8] tried to generate more images by changing the background and clothing in existing images. Although these methods improve the performance of the 3D human pose estimation model by generating more training images, the fundamental problem has not been solved since the gap between synthesized images and real images still exists.

2.2.3 Weakly-supervised Learning

Other works try to solve this problem by weakly supervised methods [13, 22, 23]. Kocabas et al. [13] introduces a self-supervised learning method to train a 3D pose model by generating 3D pose labels according to multi-view information and camera parameters. Chen et al. [22] proposes to learn a geometry representation from multi-view information. The learned representation could map to a 3D pose with a shallow network and less annotated training samples. Wandt et al. [23] proposes an adversarial reprojection network to train the 3D pose model. They predict 3D pose and camera parameters at the same time, then project the 3D pose to 2D pose by camera information. Although these methods reduce the reliance on 3D pose annotations, the performance gap between weakly supervised methods and fully supervised methods is huge. Meanwhile, camera parameters and multi-view formation can’t be provided in existing monocular images, which limits the application of weakly supervised methods.

2.3 Generalization in the wild

The above methods to solve the problems of lacking 3D pose data have alleviated the problems of generalization [15, 44, 45]. Other methods study the generalization according to different viewpoints. Fang et al. [44] utilized virtual camera simulation to generate more 2D pose data from different viewpoints and designed a generative adversarial network (GAN) to learn pose grammar, which can encode the human body configuration for 3D human pose estimation. Wang et al. [45] proposed a viewpoint prediction network to improve cross-dataset generalization for 3D human pose estimation. Zeng et al. [15] introduced a split-and-recombine approach to improve generalization by recombining rare or unseen poses in the training set. These methods try to improve the generalization by enhancing the adaptability of the model. However, they can’t handle the real images in the wild since they are designed especially for existing datasets.

3 Approach

To address the problems of lacking 3D labels and generalizing the 3D pose model to the in-the-wild scenes, we propose a framework that pre-training the 3D pose network on the 2D dataset and finetuning this network on the downstream task of 3D human pose estimation. We propose an approach to directly extract relative depth information on RGB images via perspective knowledge, which is used as the supervision of learning 3D representation for weakly-supervised pre-training (WSP). The WSP strategy enables the network to learn how to distinguish the relative depth of two points in an image, while can help 3D pose estimation. For pre-training, we collected a dataset called MCPC. We pre-trained WSP on MCPC and finetuned on 3D pose datasets.

3.1 Generating Relative Depth via Perspective Knowledge

As shown in Figure 1 (c), for an image with multiple persons, it’s easy for humans to estimate that person P1P1 is closer to the camera than the remaining people (P2,P3,P4P2,P3,P4). We can estimate the relative depth according to the perspective knowledge that objects close to the camera in the image are larger, while objects far away from the camera are smaller. Meanwhile, assuming that the 2D images are all captured by the camera at a horizontal angle on the ground, then a point on the ground in the image with a greater YY-axis value is closer to the camera than other points on the ground in the image with smaller YY-axis value. These are two of the experience of perspective knowledge. In the following, we introduce the details to generate relative depth.

For the joints (e.g., head, ankle, etc.) that belong to the persons in an image, the goal is to obtain their relative depths. Supposing that the human body is standing on the ground, we can simplify that ankle points are ground points. Then, we can obtain the relative depths of ankles belonging to different persons according to the above knowledge of human experience. For example, for ankle points PAankleP_{A}^{~{}ankle} belong to person AA and ankle points PBankleP_{B}^{~{}ankle} belong to person BB in Figure 2 (a), the relative depths between PAankleP_{A}^{~{}ankle} and PBankleP_{B}^{~{}ankle} is

ZC(PAankle)<ZC(PBankle)Z^{C}(P_{A}^{~{}ankle})<Z^{C}(P_{B}^{~{}ankle}) (1)

since

YI(PAankle)>YI(PBankle),Y^{I}(P_{A}^{~{}ankle})>Y^{I}(P_{B}^{~{}ankle})~{}, (2)

where ZC()Z^{C}(\cdot) means the depth in camera coordinate system. YI()Y^{I}(\cdot) means the YY-axis value in image coordinates system.

Now, the relative depth between person AA and person BB can be obtained according to Equation 1, that is

ZC(A)<ZC(B),Z^{C}(A)<Z^{C}(B)~{}, (3)

which means that person AA is closer to the camera than person BB. Although the all joints on person AA are not always closer to the camera than the all joints on person BB, to obtain the relative of all human joints, let’s simplify the condition that

ZC(PAj)<ZC(PBj),j=1,,J,Z^{C}(P_{A}^{j})<Z^{C}(P_{B}^{j}),~{}j=1,...,J~{}, (4)

where JJ is the number of human joints. Equation 4 means that the joints on person AA are closer to the camera than the joints on person BB.

Note that there are some negative examples when person AA is extremely close to person BB, such as AA hugs BB. Although we ignore these cases here, our experimental results show that our model has the capability to handle these incorrect labels, which can be regarded as noises.

We obtain relative depths of human joints on 2D datasets according to Equation 1 - 4, which are used as the supervision to train network.

3.2 MCPC Dataset

To tackle the problem of lacking enough 3D pose data with diversified posture and background, we want to extract weak 3D information from 2D datasets for 3D human pose estimation. There are several 2D datasets with a lot of 2D pose annotations, which are MPII [17], COCO [18], Posetrack [26] and CrowdPose [27]. However, some images in these datasets may only contain a single person or multiple persons without the annotations of the ankle. This can’t meet the requirements to generate relative depth. Therefore, we select images from these 2D datasets to form a new dataset, named MCPC.

The images of MCPC come from four 2D datasets: MPII, COCO, Posetrack, and CrowdPose. We collect images according to the two principles: 1)There are at least two persons in an image, 2)The ankle joints must be annotated in these persons. After filtering images, we collect over 53k images with more than 220k human instances. Since each image in MCPC has different human subjects with different postures and backgrounds, MCPC is a larger scale 2D dataset in natural scenes. For each image in the MCPC dataset, we generate relative depths for each human joint according to the perspective knowledge introduced in Section 3.1.

We make a comparison of several widely-used 2D datasets and 3D datasets in Table 1. Existing 3D pose datasets (e.g. Human3.6M, MuCo), are collected in indoor environments. Although there are over 300k images in these datasets, there are only 7 and 8 different human subjects in the Human3.6M and MuCo datasets, respectively. The images in these datasets have a high degree of similarity since they are captured with several subjects in the same place. The images in 2D datasets (e.g. MPII, COCO, Posetrack, and CrowdPose) are captured in different environments with different human subjects. However, they have no 3D labels. The MCPC dataset has multiple instances in an image with 2D pose labels and relative depth labels. The relative depth labels are the key information to improve the performance of 3D human pose estimation and the generalization ability of the 3D pose model.

Table 1: Comparison of 2D and 3D pose estimation datasets. “In.” means indoor and “Out.” means outdoor. “RD” means relative depth labels. Existing 3D pose estimation datasets (Human3.6M [7], Muco [9]) are collected with a few subjects performing different actions in limited environments.
Dataset Images Instances Subjects Environment Labels
Human3.6M [7] \sim312k \sim312k 7 In. 3D pose
MuCo [9] \sim400k \sim400k 8 In. 3D pose
MPII [17] \sim17k \sim28k \sim28k In./Out. 2D pose
COCO [18] \sim14k \sim56k \sim56k In./Out. 2D pose
Posetrack [26] \sim12k \sim56k \sim56k In./Out. 2D pose
CrowdPose [27] \sim20k \sim80k \sim80k In./Out. 2D pose
MCPC (Ours) \sim53k \sim220k \sim220k In./Out. 2D pose+RD

3.3 Framework

The goal of monocular 3D human pose estimation is to recover the absolute camera-centered coordinates of human joints from a 2D image. The overall architecture of our proposed framework is shown in Figure 2, which includes two steps: pre-training on 2D datasets and finetuning on 3D datasets.

Firstly, we design a network WSP to extract the keypoints features. Given an RGB image I2DI_{2D} , two-person pairs are generated by randomly selecting from multiple person instances in the image. Then, the image patches cropped from the whole image according to the bounding box of two-person pairs are sent into WSP. Inputting an image patch IABI_{AB}, which contains person instance AA and BB. The outputs of WSP are the relative depths between human joints PAP_{A} in person AA and human joints PBP_{B} in person BB. If person AA is closer to the camera than person BB, the relative depths can be denoted as Z(PAj)<Z(PBj),j[1,J]Z(P_{A}^{j})<Z(P_{B}^{j}),j\in[1,J], in which JJ is the number of human joints. Note that there may be more than two persons in an image patch because of the large pose scales of person AA and person BB. The other persons in this image patch IABI_{AB} are ignored, which means WSP just outputs the relative depths of person AA and person BB. The details of WSP and how to train WSP on 2D dataset are introduced in Section 3.4.

After pre-training WSP on 2D datasets, WSP has the ability to distinguish which joint is closer to the camera than other joints. Then, WSP is finetuned on 3D human pose estimation datasets. The pre-trained WSP can accelerate the convergence of the model for 3D human pose estimation since it has learned weak 3D representation. The pre-trained WSP also provides a good starting point to train the 3D pose model and improves the generalization ability of the model.

Refer to caption
Figure 2: The WSP framework includes (a) pre-training on MCPC dataset without 3D pose annotations and (b) finetuning on 3D pose estimation dataset. \otimes denotes element-wise multiplication. FCFC means fully connection layers. The pre-training learns better 3D representation by learning relative depth, which is beneficial for 3D pose estimation.

3.4 Relative Depth Pre-training Network

3.4.1 Model Design

WSP estimates the relative depths between human joints from a cropped image with two-person pair, which is shown in Figure LABEL:fig:f3. WSP consists of three parts: backbone, relative depth head, and 2D keypoints heatmap head.

The first one, backbone, extracts deep features from the input image by using ResNet. Based on the extracted deep features, the second part, the 2D keypoints heatmap head predict the keypoints heatmap by 3 deconvolution layers. Different from other methods to obtain 2D coordinates from 2D heatmaps, the estimated heatmaps here are used as masks to filter deep features for special human joints.

The relative depth head estimates the relative depths of human joints when given the deep features of two points. The inputs of relative depth head are deep features filtered by keypoints heatmaps. Deep features FdF_{d} has a size of (D, H, W), and 2D heatmap MM has a size of (J, H, W). D is the channel number. H and W are the height and width of heatmaps. J represents J joints. The keypoints deep features are extracted as follows

Fj=Fdrepeat(Mj,D),j=1,,JF^{j}=F_{d}\otimes repeat(M^{j},D),j=1,...,J (5)

where jj denotes jj-th joint. repeatrepeat means the operation of repeating MjM^{j} of size H×WH\times W into size D×H×WD\times H\times W. \otimes denotes element-wise multiplication. We ignore other persons who may are cropped in this image patch and their relative depths are estimated in other image patches. MjM^{j} represents the 2D heatmaps for the jj-th keypoint. FjF^{j} represents the filtered deep features for the jj-th keypoint. Then, for each person, FjF^{j} is extracted into a vector of size DD by the pooling layer within the bounding box of this person. Thus, the extracted deep features have a size of (J, D).

Then, we randomly select N(N<J)N~{}(N<J) keypoints for each person to estimate their relative depths. We use a fully connected network to classify the relative depths of selected keypoints. As shown in Figure 2, for the person AA and person BB in the input image, we firstly obtain the filtered deep feature F(PA)F(P_{A}) and F(PB)F(P_{B}) by Equation 5, which have a size of (J, D). After randomly selecting, we get Fs(PA)F_{s}(P_{A}) and Fs(PB)F_{s}(P_{B}), which have a size of (N, D). For each randomly selected keypoint, we concatenated Fsj(PA)F_{s}^{j}(P_{A}) and Fsj(PB)F_{s}^{j}(P_{B}) and send them into the fully connected network to estimate their relative depths. The predicted relative depths can be denoted as

RABj=f(Fs(PAj)Fs(PBj)),R_{AB}^{j}=f(F_{s}(P_{A}^{j})\oplus F_{s}(P_{B}^{j}))~{}, (6)

where \oplus represents concatenate operation. f()f(\cdot) represents fully connected network of input dimension 2×D2\times D and output dimension 11. RABjR_{AB}^{j} represents the estimated relative depths for jj-th joint in person AA and person BB.

3.4.2 Loss Function

Given an image patch IABI_{AB} cropped from original image I2DI_{2D}, WSP estimates the relative depths RABjR_{AB}^{j} for each joints in person AA and BB according to the Equation 5 and 6. The ground-truth of relative depths is R^ABj\hat{R}_{AB}^{j}, which can be generated according to Equation 1, 2, 3 and 4. To numerical represent relative depths, we define that

R^ABj={1,Z(PAj)Z(PBj)0,Z(PAj)>Z(PBj)\displaystyle\hat{R}_{AB}^{j}=\begin{cases}1,\quad Z(P_{A}^{j})\leq Z(P_{B}^{j})\\ 0,\quad Z(P_{A}^{j})>Z(P_{B}^{j})\end{cases} (7)

where PAjP_{A}^{j}, PBjP_{B}^{j} and Z()Z(\cdot) are the jj-th joint in person AA, the jj-th joint in person BB and the operation to get depths, respectively.

We use cross-entropy loss and L2L_{2} loss for relative depth learning and heatmap learning, respectively. The total training loss is

=hm+α×rd,\mathcal{L}=\mathcal{L}_{hm}+\alpha\times\mathcal{L}_{rd}~{}, (8)

where hm\mathcal{L}_{hm}, rd\mathcal{L}_{rd} and α\alpha are the loss for keypoints heatmap, keypoints relative depth, and loss weight, respectively.

rd\mathcal{L}_{rd} is cross-entropy loss, and

rd=1JjJ[R^jlog(Rj)+(1R^j)log(1Rj)],\mathcal{L}_{rd}=\frac{1}{J}\sum_{j}^{J}-[\hat{R}^{j}log(R^{j})+(1-\hat{R}^{j})log(1-R^{j})]~{}, (9)

where JJ, R^j\hat{R}^{j}, and RjR^{j} are the number of human joints, the ground-truth of relative depths for jj-th joint, and the relative depth estimated by WSP for jj-th joint, respectively.

hm\mathcal{L}_{hm} is L2L_{2} loss, and

hm=1JKkKjJMkjM^kj2,\mathcal{L}_{hm}=\frac{1}{JK}\sum_{k}^{K}\sum_{j}^{J}||M_{k}^{j}-\hat{M}_{k}^{j}||_{2}~{}, (10)

where JJ, KK, MkjM_{k}^{j} and M^kj\hat{M}_{k}^{j} are the number of human joints, the number of estimated persons in the input image, the predicted keypoints heatmaps, and the ground-truth of keypoints heatmaps for jj-th joint in person kk, respectively.

3.5 Finetuning WSP

3.5.1 Model design

We finetune WSP on the 3D human pose estimation dataset. For 3D human pose estimation, as [10, 13, 11], we add a head to learn 3D keypoints heatmap. The 3D coordinates can be obtained by integrating 3D heatmaps. The integrating operation to get xx values of 3D coordinates can be formulated as

Cxj=x=1Wz=1Dy=1HM^j(p),C_{x}^{j}=\sum_{x=1}^{W}\sum_{z=1}^{D}\sum_{y=1}^{H}\hat{M}^{j}(p)~{}, (11)

where pp represents the points in 3D heatmap M^\hat{M}. CxjC_{x}^{j} is the xx-axis values of 3D coordinates for jj-th joints. CyjC_{y}^{j} and CzjC_{z}^{j} are obtained also according to the Equation 11 by changing the order of integrating.

3.5.2 Loss Function

Following the suggestion from [10], we use L1L1 loss to optimize the errors of coordinates. The training loss of finetuning WSP is

3d=1JjJCjC^j1,\mathcal{L}_{3d}=\frac{1}{J}\sum_{j}^{J}||C^{j}-\hat{C}^{j}||_{1}~{}, (12)

where JJ, CjC^{j} and C^j\hat{C}^{j} are the numbers of human joints, the estimated coordinates by our network and the ground-truth coordinates, respectively.

4 Experiments

4.1 Dataset and Evaluation Metrics

Our WSP is pre-trained on MCPC dataset, then finetuned on Human3.6M and MuCo-3DHP datasets. Here, wo introduce the experimental protocol and evaluation metrics.

4.1.1 MCPC Dataset

The images in MCPC dataset for training are all collected from the train set in MPII, COCO, PoseTrack, and CrowdPose dataset. The train set of MCPC contains over 53k images with more than 220k persons. The images in the test set of MCPC are all from the test set in these four 2D datasets. The test set of MCPC contains over 11k images with more than 18k two-person pairs.

For relative depth pre-training, we evaluate the predicted relative depths of human joints on the test set of the MCPC dataset. Since it is formulated as a classification task, we use AUC, Accuracy, Precision, Recall, and F1 score to evaluate the predicted relative depths. The threshold for accuracy, precision, recall, and F1 score is 0.5.

4.1.2 Human3.6M Dataset

Human3.6M [7] is the largest and most widely-used benchmark for 3D human pose estimation. This dataset is captured by the MoCap system in the lab environment. It consists of 3.6 million videos frames from 4 cameras’ viewpoints. 11 subjects perform 15 activities, such as discussion, eating, etc. The appearance of the human body and the background of the images is simple since the limited subjects and lab environment. Accurate 3D annotations of the human pose are generated by a motion capture system in each frame.

For evaluation, we use mean per joint position error (MPJPE), which is widely used in previous works. Some works also use another widely-used metric, PA MPJPE. They firstly align the predicted 3D pose and ground truth 3D pose with a rigid transformation using Procrustes Analysis and then compute MPJPE, which is called PA MPJPE.

4.1.3 MuCo-3DHP and MuPoTS-3D Datasets

MuCo-3DHP and MuPoTS-3D Datasets are 3D multi-person pose estimation datasets proposed by Mehta et al. [9]. MuCo-3DHP is the training set, which is generated by compositing the existing MPI-INF-3DHP 3D single-person pose estimation dataset [8]. The testing set, MuPoTS-3D dataset is captured outdoors and includes 20 real-world scenes with ground-truth 3D poses for up to three subjects. The 3D annotations of the MuPoTS-3D dataset are obtained with a multi-view marker-less motion capture system.

For evaluation, 3DPCK, a 3D percentage of correct keypoints is used after alignment with ground truth. It treats a joint’s prediction as correct if it lies within 15cm from the ground-truth joint location. For more rigorous testing, we also additionally use 3DPCK with thresholds of 10cm, 11cm, 12cm, 13cm, and 14cm, respectively.

Table 2: Ablations on MCPC dataset in RD Protocol 1.
Settings ΔS\Delta S AUC(%) Acc(%) P(%) R(%) F1(%)
WSP-A \leq1.0 95.39 88.03 88.04 88.04 88.04
WSP-B \leq0.5 94.89 87.29 87.29 87.30 87.29
WSP-C \leq0.3 92.78 84.62 84.67 84.59 84.63
WSP-D \leq0.1 90.34 81.66 81.69 81.62 81.65
Table 3: Ablations on MCPC dataset in RD Protocol 2. RA means remove the ankle parts in image for testing.
Settings RA ΔS\Delta S AUC(%) Acc(%) P(%) R(%) F1(%)
WSP-A ×\times \leq1.0 95.39 88.03 88.04 88.04 88.04
WSP-E \surd \leq1.0 93.72 85.70 85.66 85.70 85.68
WSP-F \surd \leq0.5 92.92 84.63 84.64 84.63 84.63
WSP-G \surd \leq0.3 90.21 81.56 81.50 81.67 81.58
WSP-H \surd \leq0.1 87.89 79.00 78.96 79.03 79.00

4.2 Experimental Protocol

4.2.1 MCPC Dataset

We formulate the task of estimating relative depths as distinguishing which human joint is closer to the camera. When training, for each image in mini-batch, we randomly select two persons in this image to estimate their relative depths of human joints. Due to the different scales of the two-person pairs, we define that

S=PhtPhd2,S=||P_{ht}-P_{hd}||_{2}~{}, (13)

where SS denotes the scale of the person. PhtP_{ht} and PhdP_{hd} are the head_top joint and head_down joint in the person, respectively.

For person AA and person BB, the normalized scale gap ratio ΔS\Delta S is denoted as

ΔS=|SASB|max(SA,SB),\Delta S=\frac{|S_{A}-S_{B}|}{max(S_{A},~{}S_{B})}~{}, (14)

where SAS_{A}, SBS_{B} are the scales of person AA and person BB generated by Equation 13. 0ΔS10\leq\Delta S\leq 1.

In order to fully verify that WSP has learned the knowledge of estimating relative depths, we define two protocols. RD protocol 1: we evaluate the estimated relative depths in different scale gap ratios (ΔS=1.0,0.5,0.3,0.1\Delta S=1.0,0.5,0.3,0.1). The smaller ΔS\Delta S indicates that the person AA and person BB have similar scales, which means it’s more difficult to estimate their relative depths. RD Protocol 2: Removing the human ankle parts in the input image for testing. RD Protocol 2 is designed to study if WSP still can distinguish the relative depth without the ankle clues.

4.2.2 Human3.6M Dataset

Three experimental protocols are used. Following previous works [11, 10], the first one uses five subjects (S1, S5, S6, S7, S8) as training data and the other two subjects (S9, S11) as testing data. We denote it as the Subject Protocol 1, which uses MPJPE as the evaluation metric. The second one uses six subjects (S1, S5, S6, S7, S8, S9) as the training data and the other one subject (S11) as testing data. We denote it as the Subject Protocol 2, which uses PA MPJPE as the evaluation metric. To evaluate the generalization of 3D pose estimation model, we also use another widely-used protocol following [35, 15], called Cross Action Protocol, which uses MPJPE as the evaluation metric. Cross Action Protocol trains on only one of the 15 actions in the Human3.6M dataset and tests on all actions, which is designed to test the generalization ability of the 3D pose model.

4.2.3 MuCo-3DHP and MuPoTS-3D Datasets

Following the previous work [11], we use 400k frames composited from the MuCo-3DHP dataset for training, in which half of the images are background augmented. As the setting of previous works, we also use the COCO dataset for augmentation. When training models, each mini-batch consists of half MuCo-3DHP and half COCO images. We use the strategy of auxiliary training as [10, 11] since the COCO dataset has no annotations of the 3D pose.

Table 4: Comparison with baseline with no relative depth pre-training. For fair comparison, PoseNet and WSP are finetuned on Human3.6M (HM36) dataset and finetuned on mixed HM36 and MCPC dataset.
Methods Backbone Dataset MPJPE(mm)
Action Avg Δ\Delta Joints Avg Δ\Delta X Δ\Delta Y Δ\Delta Z Δ\Delta
PoseNet [11] R-50 HM36 67.5 \downarrow13.2 59.9 \downarrow12.3 22.0 \downarrow4.6 24.1 \downarrow5.4 50.2 \downarrow10
WSP(Ours) 54.3 47.6 17.4 18.7 40.2
PoseNet [11] R-50 HM36+MCPC 52.9 \downarrow5.4 47.7 \downarrow5.4 18.8 \downarrow2.5 18.3 \downarrow2.2 39.9 \downarrow4.4
WSP(Ours) 47.5 42.3 16.3 16.1 35.5

4.3 Implementation Details

4.3.1 Relative Depth Pre-training

We use ResNet pre-trained on ImageNet as the backbone of WSP. Then, we train WSP on the MCPC dataset. We summarize the hyper-parameters as follows for pre-training: the loss weight α\alpha (in Equation 8) equals 50. The input image size is 256×256256\times 256 and the heatmap size is 64×6464\times 64. The batch size is 64 and the initiating learning rate is 0.001. The learning rate is decay to 0.0001 and 0.00001 at 3030th epoch and 4040th epoch, respectively. The total epoch number for training is 50. We conduct the experiments using PyTorch on the Linux platform, and 2 V100 GPUs are used. It cost about 50 hours to train WSP on the MCPC dataset.

4.3.2 Finetuning on 3D Human Pose Estimation Dataset

We use WSP pre-trained on the MCPC dataset as the backbone network for 3D human pose estimation. We remove the original heads of WSP and add an up-sampling head to estimate the 3D human pose. We finetune the network on Human3.6M and MuCo-3DHP datasets, respectively. We summarize the hyper-parameters as follows for WSP finetuning: The input image size is 256×256256\times 256 and the heatmap size is 64×6464\times 64. The batch size is 64 and the initiating learning rate is 0.001. The learning rate is decay to 0.0001 and 0.00001 at 1717th epoch and 2121th epoch, respectively. The total epoch numbers for training is 25. We conduct the experiments using PyTorch on the Linux platform, and 2 V100 GPUs are used. It cost about 8 hours and 20 hours to finetune WSP on Human3.6M and MuCo-3DHP datasets, respectively.

Table 5: Comparison with State-of-the-art Methods on the 3D pose dataset (Human3.6M) in Subject Protocol 1 with MPJPE and Subject Protocol 2 with PA MPJPE. Bold indicates the best and underline indicates the second best
MPJPE(mm) Dir. Dis. Eat. Gre. Phoe. Pho. Pos. Pur. Sit. SiD. Smo. Wai. WaD. Wal. WaT. Avg
Jahangiri et al. [46] ICCV’17 74.4 66.7 67.9 75.2 77.3 70.6 64.5 95.6 127.3 79.6 79.1 73.4 67.4 71.8 72.8 77.6
Mehta et al. [8] 3DV’17 57.5 68.6 59.6 67.3 78.1 56.9 69.1 98.0 117.5 69.5 82.4 68.0 55.3 76.5 61.4 72.9
Martinez et al. [35] ICCV’17 51.8 56.2 58.1 59.0 69.5 55.2 58.1 74.0 94.6 62.3 78.4 59.1 49.5 65.1 52.4 62.9
Fang et al. [44] AAAI’18 50.1 54.3 57.0 57.1 66.6 53.4 55.7 72.8 88.6 60.3 73.3 57.7 47.5 62.7 50.6 60.4
Sun et al. [47] ECCV’17 52.8 54.8 54.2 54.3 61.8 53.1 53.6 71.7 86.7 61.5 67.2 53.4 47.1 61.6 63.4 59.1
Pavlakos et al. [24] CVPR’18 48.5 54.4 54.4 52.0 59.4 65.3 49.9 52.9 65.8 71.1 56.6 52.9 60.9 44.7 47.8 56.2
Sun et al. [10] ECCV’18 47.5 47.7 49.5 50.2 51.4 43.8 46.4 58.9 65.7 49.4 55.8 47.8 38.9 49.0 43.8 49.6
Moon et al. [11] ICCV’19 50.5 55.7 50.1 51.7 53.9 46.8 50.0 61.9 68.0 52.5 55.9 49.9 41.8 56.1 46.9 53.3
Cai et al. [36] ICCV’19 46.5 48.8 47.6 50.9 52.9 61.3 48.3 45.8 59.2 64.4 51.2 48.4 53.5 39.2 41.2 50.6
Zeng et al. [15] ECCV’20 - - - - - - - - - - - - - - - 49.9
Lin et al. [48] CVPR’21 - - - - - - - - - - - - - - - 54.0
Li et al. [49] ICCV’21 - - - - - - - - - - - - - - - 48.6
Wehrbein et al. [50] ICCV’21 38.5 42.6 39.9 41.7 46.5 51.6 39.9 40.8 49.5 56.8 45.3 46.4 46.8 37.8 40.4 44.3
WSP (Ours) 38.2 42.5 40.5 42.1 45.3 46.6 34.5 46.2 49.1 62.4 44.2 39.6 43.1 34.8 39.7 43.2
PA MPJPE(mm) Dir. Dis. Eat. Gre. Phoe. Pho. Pos. Pur. Sit. SiD. Smo. Wai. WaD. Wal. WaT. Avg
Moreno et al. [51] CVPR’17 67.4 63.8 87.2 73.9 71.5 69.9 65.1 71.7 98.6 81.3 93.3 74.6 76.5 77.7 74.6 76.5
Zhou et al. [16] TPAMI’18 47.9 48.8 52.7 55.0 56.8 49.0 45.5 60.8 81.1 53.7 65.5 51.6 50.4 54.8 55.9 55.3
Martinez et al. [35] ICCV’17 39.5 43.2 46.4 47.0 51.0 41.4 40.6 56.5 69.4 49.2 56.0 45.0 38.0 49.5 43.1 47.7
Sun et al. [47] ECCV’17 42.1 44.3 45.0 45.4 51.5 43.2 41.3 59.3 73.3 51.0 53.0 44.0 38.3 48.0 44.8 48.3
Fang et al. [44] AAAI’18 38.2 41.7 43.7 44.9 48.5 40.2 38.2 54.5 64.4 47.2 55.3 44.3 36.7 47.3 41.7 45.7
Sun et al. [10] ECCV’18 36.9 36.2 40.6 40.4 41.9 34.9 35.7 50.1 59.4 40.4 44.9 39.0 30.8 39.8 36.7 40.6
Cai et al. [36] ICCV’19 36.8 38.7 38.2 41.7 40.7 46.8 37.9 35.6 47.6 51.7 41.3 36.8 42.7 31.0 34.7 40.2
Moon et al. [11] ICCV’19 31.0 30.6 39.9 35.5 34.8 30.2 32.1 35.0 43.8 35.7 37.6 30.1 24.6 35.7 29.3 34.0
Wehrbein et al. [50] ICCV’21 27.9 31.4 29.7 30.2 34.9 37.1 27.3 28.2 39.0 46.1 34.2 32.3 33.6 26.1 27.5 32.4
WSP (Ours) 25.1 27.7 26.1 27.2 31.3 32.1 25.6 30.6 31.8 37.5 28.1 24.8 24.2 25.6 26.1 28.2

4.4 Ablation Study

4.4.1 Relative Depth Learning

We formulate the pre-training task as a classification task, which aims to learn to distinguish the relative depth relationship between two human joints. After training WSP on the MCPC dataset, we test WSP on the test set of the MCPC dataset. The results of classification are shown in Table 2 and Table 3.

In Table 2, WSP-A, WSP-B, WSP-C, and WSP-D represent different test settings. All of the settings are WSP with the ResNet-50 backbone. The test images are cropped from the images in the testing set of MCPC. The cropped image includes two persons, which are the targets to distinguish relative depth relationships. Since the scales of the two human bodies in the image may be very different, we define the scale gap ratio ΔS\Delta S to represent this scale gap. WSP-A, ΔS1.0\Delta S\leq 1.0 means that all the cropped images are used for testing, regardless of the huge scale gap between the two persons in this cropped image. WSP-B, ΔS0.5\Delta S\leq 0.5 means that only a part of cropped images is used for testing. The scale gap of two persons in the cropped image must be less than or equal to 0.5, which means the testing images are more difficult to distinguish relative depth relationships. With the same knowledge, WSP-C (ΔS0.3\Delta S\leq 0.3) and WSP-D (ΔS0.1\Delta S\leq 0.1) mean the smaller and more difficult testing set.

WSP-A obtains 95.39% in AUC. The values of accuracy, precision, recall, and F1 score are all over 88%. These results illustrate that WSP can distinguish the relative depth relationship of keypoints when given an image with two persons. WSP-B, WSP-C, and WSP-D are designed to study the influence of the scale. The smaller ΔS\Delta S means that the two persons in testing images are with more similar scales. As shown in the results in Table 2, all the evaluation metrics decrease with the smaller scale gap ratio ΔS\Delta S, which verifies that scale is an important visual cue to distinguish relative depth. However, although we just keep the hardest cases for evaluation in WSP-D, we still can obtain over 90% in AUC and over 81% in accuracy, precision, recall, and F1 score, respectively. The results show that WSP can estimate relative depths successfully when two persons have almost the same scales.

Due to the relative depth supervision being generated according to the ankle cues, we study the influence of ankle cues in Table 3. RA in Table 3 means removing the ankle parts in the images for testing. WSP-E, WSP-F, WSP-G, and WSP-H are designed to that WSP still can estimate relative depths successfully without the guiding of ankle clues. Although there are no ankle joints in the input images, for the testing images with almost the same scales, WSP-H still can obtain over 87% in AUC and over 79% in accuracy, precision, recall, and F1 score, respectively. The results show that the accuracy of estimating relative depths is influenced by ankle joints, but WSP can still have a good estimation of relative depths even without the clues of ankle joints.

4.4.2 3D Human Pose Estimation based on WSP

To verify that WSP is beneficial to 3D human pose estimation, we finetune WSP on the Human3.6M dataset. The baseline is PoseNet [10, 13, 11], which is a widely-used network for 3D human pose estimation. PoseNet and WSP are the same networks, the difference is that WSP is pre-trained on the MCPC dataset. As shown in Table 4, PoseNet achieves 67.5 in MPJPE of action average based on ResNet-50. WSP achieves 54.3 in MPJPE and a relative gain of 20% with the pre-trained weights.

In order to further analyze the influence of pre-training weights, we make a comparison of PoseNet and WSP on MPJPE of joints average. We can find that the MPJPE of joints averages on ZZ coordinates are obviously greater than the MPJPE of joints average on XX and YY coordinates from Table 4. Here, we denote them as MPJPEXMPJPE_{X}, MPJPEYMPJPE_{Y}, and MPJPEZMPJPE_{Z}, respectively. MPJPEZMPJPE_{Z} of PoseNet are greater than 50mm, while MPJPEXMPJPE_{X} and MPJPEYMPJPE_{Y} of PoseNet are about 23mm. The phenomenon shows that the error of depth predicting is the key problem to improve the performance of 3D human pose estimation. However, WSP obtains 40.2 in MPJPEZMPJPE_{Z} of action average based on ResNet-50. Compared with PoseNet, the MPJPEZMPJPE_{Z} decay to 40mm from 50mm, which verifies that the pre-trained WSP is beneficial to 3D pose estimation, especially in the depth prediction. We believe that the relative depth pre-training on MCPC learns how to distinguish the relative depth, which improves 3D pose estimation.

4.4.3 MTS vs WSP

Since we use the MCPC dataset for the pre-training of WSP, to verify the improvements in Section 4.4.2 come from more data or our pre-training strategy, we make the comparison of mixing training strategy (MTS) and WSP in this section.

MTS is adding 2D datasets with 3D datasets for mixing training. MTS just gives 2D supervision when using images in 2D datasets while 3D supervision when using images in 3D datasets. Following [10, 13, 11], we also finetune WSP on Human3.6M and MCPC datasets to ensure that WSP and MTS are trained on the same data for a fair comparison. The results are shown in Table 4. PoseNet achieves 52.9 MPJPE based on ResNet-50. WSP achieves 48.7 in MPJPE based on ResNet-50. Compared with PoseNet, the relative improvement is 6%. The MPJPEZMPJPE_{Z} of WSP decays to 35.5mm with ResNet-50, which has an improvement of 11% in estimating depths compared with 39.9mm of PoseNet. The MTS tries to improve the performance of predicting 3D pose by adding the constraint of the 2D pose, which to further enhance the structural constraint of the 3D pose. However, WSP tries to directly learn weak 3D representation from 2D datasets, which significantly improves the accuracy of depth prediction for 3D human pose estimation as the results are shown in Table 4.

Therefore, the WSP improves the performance of 3D pose estimation since the relative depth pre-training.

4.5 Comparison with State-of-the-art Methods

4.5.1 Human3.6M Dataset

We compare our proposed system with the state-of-the-art 3D human pose estimation methods on the Human3.6M dataset. Following the previous works [11], WSP is finetuned on Human3.6M dataset based ResNet-50. For evaluation on Subject Protocol 1, as shown in Table 5, WSP obtains an action average MPJPE of 43.2mm, which is the state-of-the-art result for fair comparison. For evaluation on Subject Protocol 2, as shown in Table 5, WSP obtains an action average PA MPJPE of 28.2mm, which is the state-of-the-art result for fair comparison.

Table 6: Comparison with state-of-the-art methods on multi-person MuPoTS-3D dataset.
3DPCK(%, 150mm) Backbone Category All Matched
Lcr-net [52] VGG16 Top-down 53.8 62.4
Lcr-net++ [53] ResNet50 70.6 74.0
HG-RCNN [54] ResNet50 71.3 74.2
HMOR [55] ResNet50 - 82.0
PoseNet [11] ResNet50 81.8 82.5
ORPM [9] ResNet50 Bottom-up 65.0 69.8
Xnect [56] SelecSLSNet 70.4 75.8
SMAP [57] Hourglass 73.5 80.5
WSP (Ours) ResNet50 Top-down 82.4 83.2
Table 7: Comparsion with baseline approach based on Resnet-18 on MuPoTS-3D dataset with multiple thresholds. Smaller threshold means stricter evaluation metric.
Methods Eval 3DPCK Thresholds
100mm 110mm 120mm 130mm 140mm 150mm
PoseNet [11] All 57.6 63.1 68.1 70.4 76.0 79.0
WSP(Ours) 59.5 65.1 69.9 71.9 77.2 80.1
Δ\Delta \uparrow 3.3% \uparrow 3.2% \uparrow 2.6% \uparrow 2.1% \uparrow 1.5% \uparrow 1.4%
PoseNet [11] Matched 58.6 64.2 69.4 73.8 77.5 80.5
WSP(Ours) 60.2 65.9 70.7 74.7 78.2 81.2
Δ\Delta \uparrow 2.7% \uparrow 2.6% \uparrow 1.8% \uparrow 1.2% \uparrow 1.0% \uparrow 0.9%

4.5.2 MuPoTS-3D Datasets

We compare our approach with state-of-the-art 3D human pose estimation methods on MuPoTS-3D Datasets in Table 6. MuCo-3DHP and MuPoTS-3D Datasets are outdoor multi-person 3D human pose estimation dataset. The evaluation metric is a 3D percentage of correct keypoints. Due to the multiple persons, the accuracy is computed for all ground-truth keypoints or ground-truth keypoints in matched persons. As shown in Table 6, WSP outperforms state-of-the-art methods in 3DPCK with a threshold of 150mm.

To further study the effectiveness of WSP, we compare the results of WSP and PoseNet [11] on MuPoTS-3D datasets with more strict thresholds (100mm, 110mm, 120mm, 130mm, 140mm, and 150mm) since 150mm is a broad threshold. The results are shown in Table 7. The testing threshold decreases to 100mm from 150mm, and the 3DPCK of PoseNet and WSP decreases. However, for all-person test settings, compared with PoseNet, the relative improvement of WSP increases to 3.3% from 1.4% with the more strict thresholds. For matched-person settings, the relative improvement of WSP increases to 2.7% from 0.9% with the more strict thresholds.

Refer to caption
Figure 3: The qualitative results of WSP with ResNet-50 backbone on COCO dataset. We train WSP on the MuCo dataset and test WSP on the COCO dataset. Note that there is no 3D annotation on the COCO dataset. WSP performs well even on these in-the-wild images with huge pose variances.
Table 8: Cross-action Evaluation on Human3.6M Dataset. Each entry shows that the model is trained only on the corresponding single action and tested on all actions.
PA MPJPE(mm) Dir. Dis. Eat. Gre. Phoe. Pho. Pos. Pur. Sit. SiD. Smo. Wai. WaD. Wal. WaT. Avg
Martinez et al. [35] 127.3 104.6 95.0 116.1 95.5 117.4 111.4 125.0 116.9 93.6 111.0 126.0 131.0 106.3 140.5 114.5
Ci et al. [39] 102.8 88.6 79.1 99.3 80.0 91.5 93.2 89.6 90.4 76.6 89.6 102.1 108.8 90.8 118.9 93.4
Zeng et al. [15] 92.3 71.4 71.8 86.4 66.8 79.1 82.5 86.6 88.9 93.4 66.1 83.0 74.4 90.0 97.8 82.0
WSP (Ours) 84.8 70.3 64.4 78.6 60.4 72.5 76.9 68.7 75.8 60.0 77.0 81.7 82.2 73.1 93.0 74.6

4.6 Evaluation of Generalization

The experimental results on Human3.6M [7] and MuPoTS-3D [8] datasets have shown the effectiveness of WSP. The muPoTS-3D dataset is collected in outdoor scenes. Thus, the improvements on the MuPoTS-3D dataset of WSP demonstrate the generalization ability of the WSP model.

To further verify the generalization ability of WSP, we make the comparison with previous works [35, 39, 15] following the settings of Cross Action Protocol. The results of cross-action evaluation are shown in Table 8. All the results in Table 8 are obtained by WSP, which is just trained on the single-action images and tested on the all-action images. WSP yields an overall improvement of 7.4mm (relative 9% improvement) over [15]. The results of the cross-action evaluation indicate the strong generalization ability of WSP obtained by the relative depth pre-training strategy.

In addition, we also predict 3D pose on COCO [18] dataset, which is a large-scale 2D pose dataset in the wild scenes. Due to no 3D annotations on the COCO dataset, we train WSP on the MuCo dataset and test WSP on the COCO dataset. The qualitative results of applying our methods to the in-the-wild images are shown in Figure 3. The results of these in-the-wild images show that the WSP model has strong generalization ability.

5 Conclusion

To solve the problems of lacking enough 3D pose annotations and generalization, we propose a novel approach, named weakly-supervised pre-training (WSP), which extracts weak 3D information from monocular images without the supervision of 3D annotations. WSP is pre-trained on the 2D pose datasets and finetuned on the 3D pose datasets. For pre-training, we propose a large-scale 2D dataset (MCPC), with the labels of relative depth generated according to perspective knowledge. After finetuning WSP on two 3D pose datasets, WSP achieves state-of-the-art results. The results on the COCO dataset show the strong generalization ability of WSP. To the best of our knowledge, this is the first weakly-supervised pre-training approach for 3D pose estimation, which extracts weak 3D information from 2D pose datasets without manual 3D annotations.

References

  • [1] D. Avola, M. Cascio, L. Cinque, G. L. Foresti, C. Massaroni, E. Rodolà, 2-d skeleton-based action recognition via two-branch stacked lstm-rnns, IEEE Transactions on Multimedia 22 (10) (2019) 2481–2496.
  • [2] V. Mazzia, S. Angarano, F. Salvetti, F. Angelini, M. Chiaberge, Action transformer: A self-attention model for short-time pose-based human action recognition, Pattern Recognition 124 (2022) 108487.
  • [3] Y.-L. Li, X. Liu, H. Lu, S. Wang, J. Liu, J. Li, C. Lu, Detailed 2d-3d joint representation for human-object interaction, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10166–10175.
  • [4] J. Wang, K. Qiu, H. Peng, J. Fu, J. Zhu, Ai coach: Deep human pose estimation and analysis for personalized athletic training assistance, in: Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 374–382.
  • [5] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
  • [6] B. Xiao, H. Wu, Y. Wei, Simple baselines for human pose estimation and tracking, in: Proceedings of the European Conference on Computer Vision, Springer, 2018, pp. 466–481.
  • [7] C. Ionescu, D. Papava, V. Olaru, C. Sminchisescu, Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments, IEEE Transactions on Pattern Analysis and Machine Intelligence 36 (7) (2013) 1325–1339.
  • [8] D. Mehta, H. Rhodin, D. Casas, P. Fua, O. Sotnychenko, W. Xu, C. Theobalt, Monocular 3d human pose estimation in the wild using improved cnn supervision, in: International Conference on 3D Vision, IEEE, 2017, pp. 506–516.
  • [9] D. Mehta, O. Sotnychenko, F. Mueller, W. Xu, S. Sridhar, G. Pons-Moll, C. Theobalt, Single-shot multi-person 3d pose estimation from monocular rgb, in: International Conference on 3D Vision, IEEE, 2018, pp. 120–130.
  • [10] X. Sun, B. Xiao, F. Wei, S. Liang, Y. Wei, Integral human pose regression, in: Proceedings of the European Conference on Computer Vision, Springer, 2018, pp. 529–545.
  • [11] G. Moon, J. Y. Chang, K. M. Lee, Camera distance-aware top-down approach for 3d multi-person pose estimation from a single rgb image, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 10133–10142.
  • [12] H. Qiu, C. Wang, J. Wang, N. Wang, W. Zeng, Cross view fusion for 3d human pose estimation, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 4342–4351.
  • [13] M. Kocabas, S. Karagoz, E. Akbas, Self-supervised learning of 3d human pose using multi-view geometry, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 1077–1086.
  • [14] K. Iskakov, E. Burkov, V. Lempitsky, Y. Malkov, Learnable triangulation of human pose, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 7718–7727.
  • [15] A. Zeng, X. Sun, F. Huang, M. Liu, Q. Xu, S. Lin, Srnet: Improving generalization in 3d human pose estimation with a split-and-recombine approach, in: Proceedings of the European Conference on Computer Vision, Springer, 2020, pp. 507–523.
  • [16] X. Zhou, Q. Huang, X. Sun, X. Xue, Y. Wei, Towards 3d human pose estimation in the wild: a weakly-supervised approach, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2017, pp. 398–407.
  • [17] M. Andriluka, L. Pishchulin, P. Gehler, B. Schiele, 2d human pose estimation: New benchmark and state of the art analysis, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 3686–3693.
  • [18] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, C. L. Zitnick, Microsoft coco: Common objects in context, in: Proceedings of the European Conference on Computer Vision, Springer, 2014, pp. 740–755.
  • [19] L. Pishchulin, A. Jain, M. Andriluka, T. Thormählen, B. Schiele, Articulated people detection and pose estimation: Reshaping the future, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, 2012, pp. 3178–3185.
  • [20] W. Chen, H. Wang, Y. Li, H. Su, Z. Wang, C. Tu, D. Lischinski, D. Cohen-Or, B. Chen, Synthesizing training images for boosting human 3d pose estimation, in: International Conference on 3D Vision, IEEE, 2016, pp. 479–488.
  • [21] G. Varol, J. Romero, X. Martin, N. Mahmood, M. J. Black, I. Laptev, C. Schmid, Learning from synthetic humans, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 109–117.
  • [22] X. Chen, K.-Y. Lin, W. Liu, C. Qian, L. Lin, Weakly-supervised discovery of geometry-aware representation for 3d human pose estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 10895–10904.
  • [23] B. Wandt, B. Rosenhahn, Repnet: Weakly supervised training of an adversarial reprojection network for 3d human pose estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 7782–7791.
  • [24] G. Pavlakos, X. Zhou, K. Daniilidis, Ordinal depth supervision for 3d human pose estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7307–7316.
  • [25] J. Yang, Y. Ma, X. Zuo, S. Wang, M. Gong, L. Cheng, 3d pose estimation and future motion prediction from 2d images, Pattern Recognition 124 (2022) 108439.
  • [26] M. Andriluka, U. Iqbal, E. Insafutdinov, L. Pishchulin, A. Milan, J. Gall, B. Schiele, Posetrack: A benchmark for human pose estimation and tracking, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5167–5176.
  • [27] J. Li, C. Wang, H. Zhu, Y. Mao, H.-S. Fang, C. Lu, Crowdpose: Efficient crowded scenes pose estimation and a new benchmark, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 10863–10872.
  • [28] A. Newell, K. Yang, J. Deng, Stacked hourglass networks for human pose estimation, in: Proceedings of the European Conference on Computer Vision, Springer, 2016, pp. 483–499.
  • [29] Y. Chen, Z. Wang, Y. Peng, Z. Zhang, G. Yu, J. Sun, Cascaded pyramid network for multi-person pose estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7103–7112.
  • [30] K. Sun, B. Xiao, D. Liu, J. Wang, Deep high-resolution representation learning for human pose estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 5693–5703.
  • [31] Z. Qiu, K. Qiu, J. Fu, D. Fu, Dgcn: Dynamic graph convolutional network for efficient multi-person pose estimation, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34, 2020, pp. 11924–11931.
  • [32] Z. Qiu, K. Qiu, J. Fu, D. Fu, Learning recurrent structure-guided attention network for multi-person pose estimation, in: 2019 IEEE International Conference on Multimedia and Expo (ICME), IEEE, 2019, pp. 418–423.
  • [33] A. Benzine, B. Luvison, Q. C. Pham, C. Achard, Single-shot 3d multi-person pose estimation in complex images, Pattern Recognition 112 (2021) 107534.
  • [34] S. Sharma, S. Huang, An end-to-end framework for unconstrained monocular 3d hand pose estimation, Pattern Recognition 115 (2021) 107892.
  • [35] J. Martinez, R. Hossain, J. Romero, J. J. Little, A simple yet effective baseline for 3d human pose estimation, in: Proceedings of the IEEE international conference on computer vision, 2017, pp. 2640–2649.
  • [36] Y. Cai, L. Ge, J. Liu, J. Cai, T.-J. Cham, J. Yuan, N. M. Thalmann, Exploiting spatial-temporal relationships for 3d pose estimation via graph convolutional networks, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 2272–2281.
  • [37] D. Pavllo, C. Feichtenhofer, D. Grangier, M. Auli, 3d human pose estimation in video with temporal convolutions and semi-supervised training, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019.
  • [38] L. Zhao, X. Peng, Y. Tian, M. Kapadia, D. N. Metaxas, Semantic graph convolutional networks for 3d human pose regression, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019.
  • [39] H. Ci, C. Wang, X. Ma, Y. Wang, Optimizing network structure for 3d human pose estimation, in: Proceedings of the European Conference on Computer Vision, 2019, pp. 2262–2271.
  • [40] Z. Qiu, Q. Yang, J. Wang, D. Fu, Dynamic graph reasoning for multi-person 3d pose estimation, in: Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 3521–3529.
  • [41] I. Habibie, W. Xu, D. Mehta, G. Pons-Moll, C. Theobalt, In the wild human pose estimation using explicit 2d features and intermediate 3d representations, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 10905–10914.
  • [42] W. Yang, W. Ouyang, X. Wang, J. Ren, H. Li, X. Wang, 3d human pose estimation in the wild by adversarial learning, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5255–5264.
  • [43] G. Rogez, C. Schmid, Mocap-guided data augmentation for 3d pose estimation in the wild, in: Advances in Neural Information Processing Systems, Vol. 29, Curran Associates, Inc., 2016, pp. 3108–3116.
  • [44] H.-S. Fang, Y. Xu, W. Wang, X. Liu, S.-C. Zhu, Learning pose grammar to encode human body configuration for 3d pose estimation, in: AAAI Conference on Artificial Intelligence, Vol. 32, 2018.
  • [45] Z. Wang, D. Shin, C. C. Fowlkes, Predicting camera viewpoint improves cross-dataset generalization for 3d human pose estimation, in: Proceedings of the European Conference on Computer Vision, Springer, 2020, pp. 523–540.
  • [46] E. Jahangiri, A. L. Yuille, Generating multiple diverse hypotheses for human 3d pose consistent with 2d joint detections, in: Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 2017, pp. 805–814.
  • [47] X. Sun, J. Shang, S. Liang, Y. Wei, Compositional human pose regression, in: Proceedings of the European Conference on Computer Vision, Springer, 2017, pp. 2602–2611.
  • [48] K. Lin, L. Wang, Z. Liu, End-to-end human pose and mesh reconstruction with transformers, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1954–1963.
  • [49] J. Li, S. Bian, A. Zeng, C. Wang, B. Pang, W. Liu, C. Lu, Human pose regression with residual log-likelihood estimation, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 11025–11034.
  • [50] T. Wehrbein, M. Rudolph, B. Rosenhahn, B. Wandt, Probabilistic monocular 3d human pose estimation with normalizing flows, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 11199–11208.
  • [51] F. Moreno-Noguer, 3d human pose estimation from a single image via distance matrix regression, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2823–2832.
  • [52] G. Rogez, P. Weinzaepfel, C. Schmid, Lcr-net: Localization-classification-regression for human pose, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3433–3441.
  • [53] G. Rogez, P. Weinzaepfel, C. Schmid, Lcr-net++: Multi-person 2d and 3d pose detection in natural images, IEEE Transactions on Pattern Analysis and Machine Intelligence 42 (5) (2019) 1146–1161.
  • [54] R. Dabral, N. B. Gundavarapu, R. Mitra, A. Sharma, G. Ramakrishnan, A. Jain, Multi-person 3d human pose estimation from monocular images, in: International Conference on 3D Vision, IEEE, 2019, pp. 405–414.
  • [55] C. Wang, J. Li, W. Liu, C. Qian, C. Lu, Hmor: Hierarchical multi-person ordinal relations for monocular multi-person 3d pose estimation, in: Proceedings of the European Conference on Computer Vision, Springer, 2020, pp. 242–259.
  • [56] D. Mehta, O. Sotnychenko, F. Mueller, W. Xu, M. Elgharib, P. Fua, H.-P. Seidel, H. Rhodin, G. Pons-Moll, C. Theobalt, Xnect: Real-time multi-person 3d human pose estimation with a single rgb camera, ACM Transactions Graphics 39 (4) (jul 2020).
  • [57] J. Zhen, Q. Fang, J. Sun, W. Liu, W. Jiang, H. Bao, X. Zhou, Smap: Single-shot multi-person absolute 3d pose estimation, in: Proceedings of the European Conference on Computer Vision, Springer, 2020, pp. 550–566.