This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

High Fidelity 3D Reconstructions with Limited Physical Views

Mosam Dabhi1    Chaoyang Wang1    Kunal Saluja2    László A. Jeni1    Ian Fasel2    Simon Lucey1,3
   1 Carnegie Mellon University    2 Apple Inc    3 The University of Adelaide
   https://sites.google.com/view/high-fidelity-3d-neural-prior
Abstract

Multi-view triangulation is the gold standard for 3D reconstruction from 2D correspondences given known calibration and sufficient views. However in practice, expensive multi-view setups – involving tens sometimes hundreds of cameras – are required in order to obtain the high fidelity 3D reconstructions necessary for many modern applications. In this paper we present a novel approach that leverages recent advances in 2D-3D lifting using neural shape priors while also enforcing multi-view equivariance. We show how our method can achieve comparable fidelity to expensive calibrated multi-view rigs using a limited (2-3) number of uncalibrated camera views.

1 Introduction

Triangulation refers to determining the location of a point in 3D space from projected 2D correspondences across multiple views. In theory, only two calibrated camera views should be necessary to accurately reconstruct the 3D position of a point. However, in practice, the effectiveness of triangulation is heavily dependent upon the accuracy of the measured 2D correspondences, baseline, and occlusions. As a result expensive and cumbersome multi-view rigs, sometimes involving hundreds of cameras and specialized hardware, are currently the method of choice to obtain high fidelity 3D reconstructions of non-rigid objects [20].

Deep learning has provided an alternate low-cost strategy by posing the 3D reconstruction problem as a supervised 2D-3D lifting problem – allowing for effective reconstructions with as little as a single view. Recently, there have been several breakthrough works – notably Deep NRSfM [51] and C3DPO [37] – allowing this problem to be treated in an unsupervised manner that requires ONLY 2D correspondences (i.e. no 3D supervision) greatly expanding the utility and generality of the approach. These unsupervised methods make up for the lack of physical views by instead leveraging large offline datasets containing 2D correspondences of the object category of interest. Unlike classical triangulation, these correspondences do not need to be rigid or even stem from the same object instance. Although achieving remarkable results, these deep learning methods to date have not been able to compete with the fidelity and accuracy of multi-view rigs that employ triangulation.

Although the methods using deep learning for single view 2D-3D lifting are of prominent research interest – we argue that multi-view consistency is still crucial for generating 3D reconstructions of high fidelity needed for many real-world applications. To this end, we propose a new multi-view NRSfM architecture that incorporates a neural shape prior while enforcing equivariant view consistency. We demonstrate that this framework is competitive with some of the most complicated multi-view capture rigs – while only requiring a modest number (2-3) of physical camera views. Our effort is the first we are aware of, that utilizes these new advances in neural shape priors for multi-view 3D reconstruction. Figure 1 presents a graphical depiction of our approach. Extensive evaluations are presented across numerous benchmarks and object categories including the human body, human hands, and monkey body. To clarify further, we are interested in a problem setup with multiple views that capture different instances of a deformable object – we deal with non-sequential (atemporal) data.

Motivation: In many problems, complex multi-camera rigs may be financially, technologically, or simply practically infeasible. Our work in this paper is motivated by the realization that the simplification of multi-view camera rigs – in terms of (i) the number of physical views, and (ii) the need for calibration – could open the door to a wide variety of applications including entertainment, neuroscience, psychology, ethology, as well as several fields of medicine [10, 23, 12, 7, 15].

Background: One of the most notable multi-view rigs for high-fidelity 3D reconstruction is the PanOptic studio [20], which contained 480480 VGA cameras, 3131 HD Cameras, and 1010 RGB+D sensors, distributed over the surface of a geodesic sphere with a 5.495.49m diameter. This setup also required specialized hardware for storage and gen-lock camera exposures and was aimed initially at human pose reconstruction. Despite its cost and complexity, the fidelity of the 3D reconstructions from PanOptic studio has motivated similar efforts across industry and academia. Of particular note is a recent effort that employed 6262 hardware synchronized cameras to capture the pose of Rhesus Macaque monkeys [4]. Other notable efforts include [22] for dogs,  [16] for human body, and [45, 11] for the human face.

Refer to caption
Figure 1: A traditional multi-view setup relies on the concept of triangulation with the assumption that the point being reconstructed is static in time – requiring a large number of physical views (i.e. cameras) to ensure a high fidelity reconstruction. Our approach utilizes multi-view 2D correspondences taken at the same (rigid) and different (non-rigid) points in time via a neural shape prior. Empirically (see the plot in top-right), we demonstrate that our approach can achieve comparable fidelity to expensive multi-view rigs using only two physical views. Blue lines depict the triangulation and proposed approaches (left vs. right, respectively) with as little as two physical views, and red lines show the corresponding 3D ground-truth.

Limitations: Classical multi-view triangulation can infer 3D structure solely from the rigid 2D correspondences stemming from the physical cameras at a single time instant. If given sufficient calibrated physical camera views, it remains the gold standard for 3D reconstruction. However, if calibration is unknown or the number of views is sparse, we argue that the proposed approach is of significant benefit. Our approach, though is limited in comparison to triangulation as it requires multi-view 2D correspondences taken at the same (rigid) and different (non-rigid) points in time during the learning/optimization process.

A strength of the proposed approach, however, is that the non-rigid 2D correspondences are treated atemporally (i.e. the temporal ordering of the non-rigid correspondences is completely ignored). This means that once the network weights of our approach have been learned – high-fidelity 3D estimates of the structure and cameras can be obtained in real-time from the very first frame. Changes in sampling rate or dynamics between training and run-time have no bearing on performance - for the same reason. Further, just like classical multi-view triangulation, our approach requires no 3D supervision and hence relies only on 2D correspondences. The proposed approach also assumes known 2D projected measurements so it does not directly leverage pixel intensities. Therefore, our approach can be integrated with any available 2D landmark image detector such as HR-Net [47], Stacked Hourglass Networks [36], Integral Pose Regression [48], and others. Finally, the camera is not assumed to be static making the proposed approach agnostic to camera movements.

2 Related Work

Multi-view approaches:

Multi-view triangulation [13] has been the method of choice in the context of large-scale complex rigs with multiple cameras [20, 4, 45, 11] for obtaining 3D reconstruction from 2D measurements. The number of views, 2D measurement noise, baseline, and occlusions bound the fidelity of these 3D reconstructions. These time-synchronized multiple physical views also come at considerable cost and effort. Recent work by Iskakov et al. [18] and others [43, 21, 49, 41] have explored how supervised learning can be used to enhance multi-view reconstruction. Similarly, work by Rhodin et al. [44] and Kacobas et al. [25] attempted to use supervised and self-supervised learning, respectively, to infer 3D geometry from a single physical camera view. An obvious drawback to these approaches is that one is required to have intimate 3D supervision of the object before deployment – a limitation that modern multi-view rigs are not faced with. None of these approaches are as general as the one we are proposing. For example, nearly all these prior works deal solely with the reconstruction of the human pose as they are heavily reliant upon peripheral 3D supervision.

2D to 3D Lifting:

NRSfM [6] aims to reconstruct the 3D structure of a deforming object from 2D correspondences observed from multiple views. While the object deformation has classically been assumed to occur in time [3, 30, 39, 54, 33, 42], the vision community has increasingly drawn attention to atemporal applications – commonly known as unsupervised 2D-3D lifting. These temporal approaches rely on the sequential motion of objects, our approach on the other hand is much more unconstrained – accepting uncalibrated atemporal 2D instances. Advances in unsupervised learning based approaches to 2D-3D lifting [27, 37] have seen significant improvements in their robustness and fidelity across a broad set of object categories and scenarios. These recent advances to date have only been applied to problems where there is only a single view (i.e. monocular) of the object at a particular point in time. Our approach is the first – to our knowledge – to leverage these advancements for 3D reconstruction when there are multi-view measurements taken at the same instance in time.

3 Preliminary

Notations

This paper uses the following notations throughout the manuscript.

Variable type Examples
Scalar s,N,K,Ls,N,K,L
Vector 𝐬,𝝍,𝝀\mathbf{s},\boldsymbol{\psi},\boldsymbol{\lambda}
Matrix 𝐖,𝐒,𝐑,𝐭,𝐃\mathbf{W},\mathbf{S},\mathbf{R},\mathbf{t},\mathbf{D}
Function 𝒇𝒆,𝒇𝒅,𝒈\boldsymbol{f_{e}},\boldsymbol{f_{d}},\boldsymbol{g}
lthl^{th} layer 𝝍l,l𝐃\quad{}^{l}\boldsymbol{\psi},\quad^{l}\mathbf{D}, 𝝀l\quad{}^{l}\boldsymbol{\lambda}
nthn^{th} instance 𝐖(n),𝐒(n),𝝍(n),𝝀(n)\mathbf{W}^{(n)},\mathbf{S}^{(n)},\boldsymbol{\psi}^{(n)},\boldsymbol{\lambda}^{(n)}
kthk^{th} view 𝐖k,𝐑k,𝝍k\mathbf{W}_{k},\mathbf{R}_{k},\boldsymbol{\psi}_{k}

Any different signs utilized to explain a mathematical phenomenon other than the ones described above would be explicitly defined wherever deemed necessary.

Refer to caption
Figure 2: Two views statistical shape prior. The 3D structure 𝐒\mathbf{S} is drawn from a statistical shape distribution using neural shape priors and consequently projected to 22 views using the cameras 𝐑k\mathbf{R}_{k}^{*} k[1,2]\forall k\in[1,2] – calculated through OnP formulation [46]. The proposed approach minimizes the 2D projection error between the predicted 2D projections 𝐖~k\tilde{\mathbf{W}}_{k} and target (input) 2D projections 𝐖k\mathbf{W}_{k}.

Problem setup.

We are interested in a camera rig setup with KK synchronized views capturing NN instances (samples) of non-rigid objects from the same category. Specifically, we are given a non-sequential (atemporal) dataset containing NN multi-view 2D observations {𝐖1(1),,𝐖1(N);;𝐖K(1),,𝐖K(N)}\{\mathbf{W}_{1}^{(1)},\ldots,\mathbf{W}_{1}^{(N)};\cdots;\mathbf{W}_{K}^{(1)},\ldots,\mathbf{W}_{K}^{(N)}\}, where each 𝐖P×2\mathbf{W}\in\mathbb{R}^{P\times 2} represents 2D location for PP keypoints. We want to reconstruct the 3D shape 𝐒(1),,𝐒(N)\mathbf{S}^{(1)},\ldots,\mathbf{S}^{(N)}, where each 𝐒P×3\mathbf{S}\in\mathbb{R}^{P\times 3} for each of the NN instances of the object.

Weak perspective projection.

We assume weak perspective projections, i.e. for a 3D structure 𝐒\mathbf{S} defined at a canonical frame, its 2D projection is approximated as

𝐖s𝐒𝐑xy+𝐭xy\displaystyle\mathbf{W}\approx s\mathbf{S}\mathbf{R}_{xy}+\mathbf{t}_{xy} (1)

where 𝐑xy3×2\mathbf{R}_{xy}\in\mathbb{R}^{3\times 2}, 𝐭xy2\mathbf{t}_{xy}\in\mathbb{R}^{2} are the xx-yy component of a rigid transformation, and s>0s>0 is the scaling factor inversely proportional to the object depth if the true camera model is pin-hole. If all 2D points are visible and centered, 𝐭xy\mathbf{t}_{xy} can be omitted by assuming the origin of the canonical frame is at the center of the object. Due to the bilinear form of (1), ss is ambiguous and becomes up-to-scale recoverable only when 𝐒\mathbf{S} is assumed to follow certain prior statistics. We handle scale by approximating with an orthogonal projection and solving an Orthogonal-N-Point (OnP) problem [46] to find the camera pose along with the scale, as discussed in Sec. 4.2.

Statistical shape model.

We assume a linear model for the 3D shapes 𝐒\mathbf{S} to be reconstructed, i.e. at canonical coordinates, the vectorization of 𝐒\mathbf{S} in Eq. (1), denoted 𝐬=vec(𝐒)3P\mathbf{s}=\text{vec}(\mathbf{S})\in\mathbb{R}^{3P} can be written as

𝐬=𝐃𝝍\displaystyle\mathbf{s}=\mathbf{D}\boldsymbol{\psi} (2)

where 𝐃3P×B\mathbf{D}\in\mathbb{R}^{3P\times B} is the shape dictionary with BB basis and 𝝍B\boldsymbol{\psi}\in\mathbb{R}^{B} is the code vector - taking insight from classical sparse dictionary learning methods. The factorization of 𝐒\mathbf{S} in Eq. (2) is ill-posed by nature; in order to resolve the ambiguities in this factorization, additional priors are necessary to guarantee the uniqueness of the solution. Notable priors include the assumption of 𝐒\mathbf{S} being (i)(i) low rank [8, 6, 2, 9, 31], (ii)(ii)lying in a union-of-subspaces [32, 53, 1] (iii)(iii) or compressible [26, 52, 28]. The low-rank assumption becomes infeasible when the data exhibits complex shape variations, the Union-of-subspaces NRSfM methods have difficulty clustering shape deformations and estimating affinity matrices effectively just from 2D observations. Finally, the sparsity prior allows more powerful modeling of shape variations with a large number of subspaces but suffers from sensitivity to noise.

Neural Shape Prior

Our neural shape prior is an approximation to a hierarchical sparsity prior introduced by Kong et al. [27], where each non-rigid shape is represented by a sequence of hierarchical dictionaries and corresponding sparse codes. Other neural shape priors – such as C3PDO [37] – could be entertained as well but we chose to employ Kong et al.’s method due to its simplicity with respect to enforcing multi-view equivariant constraints. The approach in [27] maintains the robustness of sparse code recovery by utilizing overcomplete dictionaries to model highly deformable objects consisting of large-scale shape variation. Moreover, if the subsequent dictionaries in this multi-layered representation are learned properly, they can serve as a filter such that only functional subspaces remain and the redundant are removed. Due to the introduction of multiple levels of dictionaries and codes in the following section, we will abuse the notation of 𝐃,𝝍\mathbf{D},\boldsymbol{\psi} by adding left superscript 11, i.e. 𝐃1,1𝝍\quad{}^{1}\mathbf{D},\quad^{1}\boldsymbol{\psi} indicating that they form the first level of hierarchy. Assuming the canonical 3D shapes are compressible via multi-layered sparse coding with lLl\in L layers, the shape code 𝝍1{}^{1}\boldsymbol{\psi} is constrained as

𝐬=1𝐃1𝝍𝝍1=2𝐃2𝝍𝝍L1=L𝐃L𝝍s.t.  l𝝍1l𝝀 , l𝝍𝟎 , l{1,,L}\displaystyle\begin{split}\mathbf{s}&=\quad^{1}\mathbf{D}\quad^{1}\boldsymbol{\psi}\\ {}^{1}\boldsymbol{\psi}&=\quad^{2}\mathbf{D}^{2}\boldsymbol{\psi}\\ &\vdots\\ {}^{L-1}\boldsymbol{\psi}&=\quad^{L}\mathbf{D}^{L}\boldsymbol{\psi}\\ \text{s.t. }\text{ }\|^{l}\boldsymbol{\psi}\|_{1}\leq\quad^{l}\boldsymbol{\lambda}&\text{ , }^{l}\boldsymbol{\psi}\geq\mathbf{0}\text{ , }\forall l\in\{1,\ldots,L\}\end{split} (3)

where 𝐃lBl1×lB{}^{l}\mathbf{D}\in\mathbb{R}^{\quad{}^{l-1}B\times\quad^{l}B} are the hierarchical dictionaries, ll is the index of hierarchy level, and 𝝀l{}^{l}\boldsymbol{\lambda} is the scalar specifying the amount of sparsity in each level. Thus, the learnable parameters are Θ={,l𝐃,l𝝀,}\Theta=\{\cdots,\quad^{l}\mathbf{D},\quad^{l}\boldsymbol{\lambda},\cdots\}. The single set of parameters Θ\Theta are fit jointly along with the sparse codes, rotation matrices, and structures 𝐒\mathbf{S} for each instance in the dataset. Jointly constraining each instance via a common set of weights (the “neural prior”) makes this work more akin to classic factorization methods, in which both the shared factors and the weightings for each instance are jointly inferred, rather than to network training approaches which aim to find weights that generalize well when later used to perform inference on unseen data.

Factorization-based NRSfM.

Equivalently, the linear model in Eq. (2) could be rewritten as

𝐒=𝐃#(𝝍𝐈3)\displaystyle\mathbf{S}=\mathbf{D}^{\#}(\boldsymbol{\psi}\otimes\mathbf{I}_{3})

where 𝐃#P×3B\mathbf{D}^{\#}\in\mathbb{R}^{P\times 3B} is a reshape of 𝐃\mathbf{D} and \otimes denotes a Kronecker product. Applying the camera matrix 𝐑xy\mathbf{R}_{xy} gives the 2D pose. Thus

𝐒𝐑xy=𝐃#(𝝍𝐑xy)\displaystyle\mathbf{S}\mathbf{R}_{xy}=\mathbf{D}^{\#}(\boldsymbol{\psi}\otimes\mathbf{R}_{xy})

Substituting the input 2D pose 𝐖\mathbf{W} from Eq. (1), we have

𝐖=𝐃#𝚿xys.t.  𝚿xy=𝝍𝐑xy and 𝝍𝒞\displaystyle\begin{split}\mathbf{W}&=\mathbf{D}^{\#}\boldsymbol{\Psi}_{xy}\\ \text{s.t. }\text{ }\boldsymbol{\Psi}_{xy}&=\boldsymbol{\psi}\otimes\mathbf{R}_{xy}\text{ and }\boldsymbol{\psi}\in\mathcal{C}\end{split} (4)

where 𝚿xy3B×2\boldsymbol{\Psi}_{{xy}}\in\mathbb{R}^{3B\times 2} is the sparse block code denoting the first two columns of 𝚿3B×3\boldsymbol{\Psi}\in\mathbb{R}^{3B\times 3}; and 𝒞\mathcal{C} denotes the neural shape prior constraints applied on the code 𝝍\boldsymbol{\psi}, e.g. hierarchical sparsity [27] in our case. Conceptually, 𝚿\boldsymbol{\Psi} is a matrix with rotations and sparse code built into it. Under the unsupervised settings, 𝐃,𝝍,𝐑,𝐒\mathbf{D},\boldsymbol{\psi},\mathbf{R},\mathbf{S} are all unknowns and are solved under the simplified assumptions that the input 2D poses are obtained through a weak perspective camera projection. We also analytically compute 𝐑\mathbf{R}^{*} as a solution to a Orthographic-n-point (OnP) problem that acts as a supervisory signal to 𝐑\mathbf{R}. Corresponding proof for 𝐑\mathbf{R}^{*} is discussed in the supplementary section.

4 Approach

Refer to caption
Figure 3: Architecture showing a 22-view 3D reconstruction approach (easily extensible to K>2K>2 views). The 2D projections from both views 𝐖k\mathbf{W}_{k} k[1,2]\forall k\in[1,2] acts as an input to encoder 𝒇e\boldsymbol{f}_{e} that extracts the block sparse code 𝚿k\boldsymbol{\Psi}_{k} from the corresponding views. A Rotation Factorization (RF) layer at the bottleneck stage shown in green, factorizes the block sparse code into the respective camera matrix 𝐑k\mathbf{R}_{k} and the unrotated vector sparse code 𝝍kL{}^{L}\boldsymbol{\psi}_{k}. The codes are then fused via pooling function 𝒈\boldsymbol{g} into a single code 𝝍L{}^{L}\boldsymbol{\psi} that acts as an input to the shape decoder 𝒇d\boldsymbol{f}_{d}. The shape decoder predicts the 3D structure 𝐒\mathbf{S} in the canonical frame while enforcing equivariant view consistency.

4.1 Bilevel optimization

Given only the input 2D poses in Eq. (4), two problems remain to address

  • How to formulate an optimization strategy to recover 𝐃,𝝍,𝐑,𝐒\mathbf{D},\boldsymbol{\psi},\mathbf{R},\mathbf{S}?

  • How to efficiently pool in KK different camera views and enforce equivariance over the predicted KK camera matrices and a single 3D structure in canonical frame?

We choose to impose neural shape priors through hierarchical sparsity constraints [27] literature for approaching a solution to the above problems, with learnable parameters Θ\Theta (see Sec. 4.2). From Eq. (4), the learning strategy of multi-view NRSfM problem for NN instances with KK views is then interpreted as solving the following bilevel optimization problem. Eq. (3) leads to relaxation of the following lower-level problem

min𝐃,Θk=1Kn=1N(min𝝍k(n)l,𝐑k(n)𝐖k(n)1𝐃(1𝚿k(n))F+l=1L𝝀ll𝚿k(n)F(3×2)+l=2L(l1𝚿k(n))l𝐃(l𝚿k(n))F)\displaystyle\begin{split}&\min_{\mathbf{D},\Theta}\sum_{k=1}^{K}\sum_{n=1}^{N}\Bigg{(}\min_{{}^{l}\boldsymbol{\psi}_{k}^{{(n)}},\mathbf{R}_{k}^{(n)}}\|\mathbf{W}_{k}^{(n)}-\quad^{1}\mathbf{D}\big{(}\quad^{1}\boldsymbol{\Psi}_{k}^{{(n)}}\big{)}\|_{F}+\\ &\sum_{l=1}^{L}\quad{}^{l}\boldsymbol{\lambda}\|^{l}\boldsymbol{\Psi}_{k}^{{(n)}}\|_{F}^{(3\times 2)}+\sum_{l=2}^{L}\|(^{l-1}\boldsymbol{\Psi}_{k}^{{(n)}})-\quad^{l}\mathbf{D}\quad(^{l}\boldsymbol{\Psi}_{k}^{{(n)}})\|_{F}\Bigg{)}\end{split} (5)

where the first expression in (5) minimizes the 2D projection error, the second expression enforces sparsity, and the third expression fits each dictionary in the hierarchy to the dictionary representation in the preceding layer. Minimizing the block Frobenius norm of 𝚿\boldsymbol{\Psi} is equivalent to minimizing the L1L_{1} norm of the vector sparse code 𝝍\boldsymbol{\psi} because 𝝍1=12𝚿F(3×2)\|\boldsymbol{\psi}\|_{1}=\frac{1}{\sqrt{2}}\|\boldsymbol{\Psi}\|_{F}^{(3\times 2)}.

4.2 Network approximate solution

The optimization problem in Eq. (5) is an instance of dictionary learning problem with sparse codes 𝝍\boldsymbol{\psi}. The classical approach to this problem is by solving the Iterative Shrinkage and Thresholding Algorithm (ISTA) [5]. However, Papyan et al. [38] show that a single layer feedforward network with Rectified Linear Unit (ReLU) activations approximate one step of ISTA, with the bias terms 𝝀l{}^{l}\boldsymbol{\lambda} adjusting the sparsity of recovered code for the lthl^{\text{th}} layer. Furthermore, the dictionaries [1𝐃,,L𝐃][\quad^{1}\mathbf{D},\ldots,\quad^{L}\mathbf{D}] can be learned by back-propagating through the feedforward network. We devise a network architecture that serves as an approximate solver to the above optimization problem and provide derivations in the following subsections.

Approximating sparse codes.

We review the sparse dictionary learning problem and consider the single-layer case stated above. To reconstruct an input signal 𝐗\mathbf{X}, the optimization problem becomes

min𝚿𝐗𝐃𝚿F+𝝀𝚿F\displaystyle\min_{\boldsymbol{\Psi}}\|\mathbf{X}-\mathbf{D}\boldsymbol{\Psi}\|_{F}+\boldsymbol{\lambda}\|\boldsymbol{\Psi}\|_{F}

As stated above, Papyan et al. [38] propose that one iteration of ISTA gives back the block-sparse codes 𝚿\boldsymbol{\Psi} as

𝚿=ReLU(𝐃𝐗;𝝀)\displaystyle\boldsymbol{\Psi}=\text{ReLU}(\mathbf{D}^{\top}\mathbf{X};\boldsymbol{\lambda})

We interpret ReLU as solving for the block-sparse code and incorporate ReLU as the nonlinearity in our encoder part of the network.

Encoder architecture.

We propose to devise an encoder network fef_{e} that takes the 2D poses as input and outputs the block sparse codes 𝚿\boldsymbol{\Psi} that has within itself the rotation matrix 𝐑\mathbf{R} as well as a rotationally invariant sparse code 𝝍\boldsymbol{\psi}, i.e. fe(𝐖k(n))(L𝚿k)f_{e}(\mathbf{W}_{k}^{(n)})\mapsto\Big{(}\quad^{L}\boldsymbol{\Psi}_{k}^{*}\Big{)}. Unrolling one iteration of ISTA for each layer, fef_{e} takes 𝐖k(n)\mathbf{W}_{k}^{(n)} as 2D pose inputs and produces block sparse codes for the last layer [1𝚿k(n),,L𝚿k(n)][\quad^{1}\boldsymbol{\Psi}_{k}^{(n)},\ldots,\quad^{L}\boldsymbol{\Psi}_{k}^{(n)}] as output, shown in Fig. 3

𝚿k(n)1=ReLU([(1𝐃#)𝐖k(n)]3×2;1𝝀(n))𝚿k(n)2=ReLU((2𝐃𝐈3)1𝚿k(n);2𝝀(n))𝚿k(n)L=ReLU((L𝐃𝐈3)L1𝚿k(n);L𝝀(n))\displaystyle\begin{split}{}^{1}\boldsymbol{\Psi}_{k}^{{(n)}}&=\text{ReLU}\Big{(}\big{[}(\quad^{1}\mathbf{D}^{\#})^{\top}\cdot\mathbf{W}_{k}^{(n)}\big{]}_{3\times 2};\quad^{1}\boldsymbol{\lambda}^{{(n)}}\Big{)}\\ {}^{2}\boldsymbol{\Psi}_{k}^{{(n)}}&=\text{ReLU}\Big{(}(\quad^{2}\mathbf{D}\otimes\mathbf{I}_{3})^{\top}\cdot\quad^{1}\boldsymbol{\Psi}_{k}^{{(n)}};\quad^{2}\boldsymbol{\lambda}^{{(n)}}\Big{)}\\ &\vdots\\ {}^{L}\boldsymbol{\Psi}_{k}^{{(n)}}&=\text{ReLU}\Big{(}\quad(^{L}\mathbf{D}\otimes\mathbf{I}_{3})^{\top}\cdot\quad^{L-1}\boldsymbol{\Psi}_{k}^{{(n)}};\quad^{L}\boldsymbol{\lambda}^{{(n)}}\Big{)}\end{split} (6)

where 𝝀(n)l\quad{}^{l}\boldsymbol{\lambda}^{{(n)}} is the learnable threshold for each layer. (l𝐃𝐈3)l1𝝍k(n)(\quad^{l}\mathbf{D}\otimes\mathbf{I}_{3})^{\top}\cdot\quad^{l-1}\boldsymbol{\psi}_{k}^{{(n)}} is implemented by a convolution transpose.

Rotation Factorization layer.

At the bottleneck, our encoder network generates a block sparse code for KK-views 𝚿k(n)L\quad{}^{L}\boldsymbol{\Psi}_{k}^{(n)}. As evident in Eq. (4), since the block sparse code has rotations 𝐑k(n)\mathbf{R}_{k}^{(n)} as well as an unrotated sparse code 𝝍k(n)\boldsymbol{\psi}_{k}^{(n)}, we add a fully-connected layer that factorizes out these quantities, named Rotation Factorization (RF) layer, shown as a green block in Fig. 3. Consequently, 𝚿k(n)L{}^{L}\boldsymbol{\Psi}_{k}^{(n)} is then factorized into an unrotated sparse code 𝝍k(n)L{}^{L}\boldsymbol{\psi}_{k}^{(n)} and the rotation matrix 𝐑k(n)\mathbf{R}_{k}^{(n)} (constraining to SO(3)SO(3) using SVD) using this fully-connected RF layer. At this stage, we pool the features from all the rotationally invariant or unrotated sparse codes 𝝍k(n)L{}^{L}\boldsymbol{\psi}_{k}^{(n)} using a sum pooling operation 𝒈\boldsymbol{g} that enforces the equivariance consistency within all the views by combining features from multiple views.

𝒈(L𝝍1(n),,L𝝍K(n))(L𝝍(n))\displaystyle\boldsymbol{g}(^{L}\boldsymbol{\psi}_{1}^{{(n)}},\ldots,^{L}\boldsymbol{\psi}_{K}^{{(n)}})\mapsto(^{L}\boldsymbol{\psi}^{{(n)}}) (7)

as shown in architecture overview Fig. 3, where 𝒈\boldsymbol{g} denotes a sum operation. Since the pooled sparse code 𝝍L{}^{L}\boldsymbol{\psi} is rotationally invariant, we generate a single canonical 3D structure 𝐒\mathbf{S} through a decoder network 𝒇d\boldsymbol{f}_{d}, that remains equivariant to KK camera rotations 𝐑1,,𝐑K\mathbf{R}_{1},\ldots,\mathbf{R}_{K}. Thus, the decoder network 𝒇d\boldsymbol{f}_{d} helps supervise the fully-connected RF layer.

Insight behind multi-view consistency.

For each individual view, we get a block sparse code representation 𝚿k(n)\boldsymbol{\Psi}_{k}^{(n)} that has the rotation 𝐑k(n)\mathbf{R}_{k}^{(n)} combined with an unrotated sparse code 𝝍k(n)\boldsymbol{\psi}_{k}^{(n)}. RF layer disentangles these quantities and generates codes that are consistent with an unrotated or canonicalized view. This architecture thus enforces equivariance consistency by consequently passing the unrotated sparse code 𝝍\boldsymbol{\psi} through a shape decoder to produce a canonicalized 3D structure. When we jointly encode multiple views into a single canonical shape, the equivariance is implicitly enforced after projecting them through the given multiple cameras. These multi-view projections help supervise the multi-view NRSfM network.

Decoder architecture.

Finally, a decoder 𝒇d\boldsymbol{f}_{d} is devised that takes input a pooled bottleneck sparse code (see Eq. 7) and generates a canonical 3D structure 𝐒\mathbf{S}. Thus, fd(L𝝍(n))(𝐒(n))f_{d}(\quad^{L}\boldsymbol{\psi}^{(n)})\mapsto\big{(}\mathbf{S}^{(n)}\big{)}

𝝍(n)L1{}^{L-1}\boldsymbol{\psi}^{(n)} =ReLU(L𝐃L𝝍(n);L𝝀(n))\displaystyle=\text{ReLU}(\quad^{L}\mathbf{D}\cdot\quad^{L}\boldsymbol{\psi}^{(n)};\quad^{L}\boldsymbol{\lambda}^{(n)})
\displaystyle\vdots
𝝍(n)1{}^{1}\boldsymbol{\psi}^{(n)} =ReLU(2𝐃2𝝍(n);2𝝀(n))\displaystyle=\text{ReLU}(\quad^{2}\mathbf{D}\cdot\quad^{2}\boldsymbol{\psi}^{(n)};\quad^{2}\boldsymbol{\lambda}^{(n)})
𝐒(n)\displaystyle\mathbf{S}^{(n)} =1𝐃1𝝍(n)\displaystyle=\quad^{1}\mathbf{D}\cdot\quad^{1}\boldsymbol{\psi}^{(n)}\vspace{-0.5em} (8)

We analytically compute a closed-form solution to 𝐑\mathbf{R}^{*} as a solution to an Orthographic-n-Point (OnP) problem that implicitly acts as supervisory signal for the 𝐑k(n)\mathbf{R}_{k}^{(n)}. Detailed proof is shown in the supplementary section.

Calculating 𝐑\mathbf{R} using solution from OnP

We are using a closed-form algebraic solution to 𝐑\mathbf{R}^{*} that gives us an optimal solution for 𝐑\mathbf{R} produced by the network at the bottleneck stage. We opt to use an algebraic solution that can be implemented as a differentiable operator and could be easily accomplished via modern autograde packages. The detailed proof for the OnP solution is given in the appendix.

Loss function

To reemphasize the loss function in our neural architecture, the loss function driving the proposed approach is a reprojection error

=1KNk=1Kn=1N𝐖k(n)𝐒(n)𝐑k(n)F\displaystyle\mathcal{L}=\frac{1}{KN}\sum_{k=1}^{K}\sum_{n=1}^{N}\|\mathbf{W}_{k}^{(n)}-\mathbf{S}^{(n)}\mathbf{R}_{k}^{(n)}\|_{F} (9)

5 Experiments

Evaluation over objects such as human body, hands, and monkey body is divided into two major categories: (i) Multi-view 3D reconstruction of an input 2D dataset, and (ii) Generation of 3D labels for unseen 2D data. The former compares against classical algorithms to generate high-fidelity 3D reconstruction from multi-view 2D input datasets. The latter discusses the generalization capability of our approach and shows that it does not overfit. For this, we follow one of the standard protocols for a human pose dataset and show results on the validation split. Equally competent 3D reconstructions are obtained for squishy deformable categories such as balloon deflation or paper tearing [19] – making the proposed approach agnostic to deformable object categories. Results from NRSfM Challenge dataset [19] using 2 camera views are shown in the supplementary section.

Network architecture and implementation details

The same neural prior architecture is used for all the KK views across different datasets. We use KK-encoders and a single shape decoder to generate one 3D structure in a canonicalized frame. The dictionary size (i.e. neural units) within each layer of encoder is decreased exponentially: {1024,512,256,128,64,32,16,8}\{1024,512,256,128,64,32,16,8\}. Ideally, if a validation set with 3D groundtruth is provided, we could select optimal architecture based on cross-validation. However, due to the unsupervised setting, we rather set the hyperparameters heuristically. For the encoder and decoder architecture discussed in Eq. (6), (8), we use a convolutional network as in Kong et al. [27] and share the convolution kernels (i.e. dictionaries) between the encoder and decoder.

Training details

We keep the same weightings for the reprojection error shown in loss function Eq. (9). We use the Adam optimizer [24] in our implementation.

Evaluation metrics

Unless otherwise noted, we utilize the following metrics to assess the prediction accuracy of 3D reconstruction. PA-MPJPE: prior to computing the mean per-joint position error, we standardize the scale of the predictions by normalizing them to match against the given ground-truth (GT) followed by rigidly aligning these predictions to GT using Procrustes alignment. Lower the better. PCK: percentage of correct keypoints after Procrustes alignment. The predicted joint is viewed as correct if the separation between the predicted and the GT joint is within a specific range (usually in cm or mm).

Monkey body dataset

OpenMonkeyStudio [4] is a huge Rhesus Macaque monkey pose dataset in a setup similar to PanOptic Studio where 6262 cameras capture the markerless pose of Rhesus Macaque monkeys. We use the provided 2D annotations over the Batch (7, 9, 9a, 9b, 10, and 11). This dataset also provides the groundtruth 3D labels for the given batches to evaluate the 3D reconstruction performance.

Human body dataset

Human 3.6 Million (H3.6M) [16] is a large-scale human pose dataset with images featuring actors performing daily activities from 4 camera views - annotated by motion capture systems. The 2D keypoint annotations of H3.6M preserve the perspective effect, and thus is a realistic dataset for evaluating the practical usage of generating 3D labels for unseen data as well as test the generalization capability of our approach. The results obtained on this dataset supports our hypothesis that the weak-perspective is a reasonable preliminary assumption; we plan to account for perspective effects as part of future work. We use this dataset for both quantitative and qualitative evaluation. For generating 3D reconstruction of an input dataset (see Sec. 5.1), we pick 5 subjects (1, 5, 6, 7, 8) and compare against the classical multi-view triangulation baselines. For generating 3D labels over unseen 2D data to showcase the generalization capability (see Sec. 5.2), we follow the standard protocol on H3.6M and use the subjects (1, 5, 6, 7, 8) during the training stage and the subjects (9, 11) for evaluation stage. Evaluation is performed on every 64th frame of the test set. We include average errors for each method.

Human hands dataset

Finally, we use an open-source hands dataset - FreiHand [55] - a large-scale open-source dataset with varied movements of hands with 3D pose annotated by motion capture systems. It consists of 3256032560 samples with their corresponding camera intrinsics. We generate random camera extrinsics and randomly create multiple camera views to generate multi-view 2D inputs for evaluating the proposed approach.

5.1 3D reconstruction of an input 2D dataset

Like multi-view triangulation or bundle adjustment, our approach jointly infers the unknown 3D shape and camera rotations from 2D keypoints. By simultaneously fitting the shared network parameters used to recover shape and pose, our approach constrains the possible reconstructions much more strongly than multi-view triangulation or bundle-adjustment approaches. Although we showcase the generalization capability of the setup by applying the fitted network to generate 3D labels for unseen 2D data (see Sec. 5.2), the major contribution of our approach is the optimization process for multi-view 3D reconstruction of an input 2D dataset. The goal is to evaluate the robustness of the proposed multi-view neural shape prior across different shape variations and hence as part of the evaluation, we report how well our method is able to reconstruct different datasets compared to the baseline methods.

Baseline

We use an implementation of iterative multi-view triangulation with robust outlier rejection [14, 13], referred to as TRNG – a method of choice for multi-view 3D human pose learning by Kocabas et al. [25]. They also provide an open-source implementation for this baseline method. A more recent method doing classical optimization on triangulation is proposed by Lee and Civera [34], however, their method is not necessarily optimal in terms of accuracy, but more in terms of computation time. TRNG first finds the points which minimize the distance from all the rays and removes the rays which are the furthest away from that point. It then re-evaluates the triangulation and this iteration is repeated 2-3 times. Empirically, we find that increasing the iteration leads us to predict near-perfect 3D reconstruction if we have exact camera calibration parameters and exact, clean 2D projections. We consider this to be a very strong baseline comparison since this approach is being widely used in industry as well as academia to generate accurate 3D reconstructions used to train 3D regression methods. We evaluate our approach on the above three datasets with substantial non-rigid deformities. For all the given experiments the cameras are chosen at random and the same set of cameras are used in the comparative baselines for a fair comparison.

Method Batch#7\#7 Batch#9\#9 Batch#9a\#9a Batch#9b\#9b Batch#10\#10 Batch#11\#11
TRNG 21.2121.21 24.3224.32 30.6730.67 24.5024.50 26.1026.10 22.7722.77
MV NRSfM 8.36\mathbf{8.36} 8.25\mathbf{8.25} 9.12\mathbf{9.12} 11.52\mathbf{11.52} 8.203\mathbf{8.203} 8.17\mathbf{8.17}
Table 1: PA-MPJPE error values for Monkey body dataset shows substantial improvement over the baseline multi-view triangulation approach while using only two views over noisy 2D keypoints. PA-MPJPE values are in cm.
S1, S5, S6, S7, S8
Extrinsics Noise Intrinsics Noise 2D keypoints Noise
σ=0.1\sigma=0.1 σ=0.5\sigma=0.5 σ=0.9\sigma=0.9 σ=0.1\sigma=0.1 σ=0.5\sigma=0.5 σ=0.9\sigma=0.9 σ=15\sigma=15 σ=25\sigma=25 σ=35\sigma=35
TRNG 65.4965.49 131.66131.66 145.94145.94 69.5769.57 188.63188.63 234.47234.47 70.0870.08 114.06114.06 154.41154.41
2-Views (ours) 30.53\mathbf{30.53} 54.22\mathbf{54.22} 65.74\mathbf{65.74} 77.82\mathbf{77.82}
Table 2: Robustness to camera calibration and 2D annotations noise for Human 3.6M dataset. Values are in mm.
Method PCK
2 Views [4] 1.2%1.2\%
4 Views [4] 59%59\%
8 Views [4] 80%80\%
16 Views [4] 82%82\%
32 Views [4] 87%87\%
48 Views [4] 95%95\%
2-views (ours) 68.63%\mathbf{68.63}\%
3-views (ours) 84.63%\mathbf{84.63}\%
Table 3: Percentage of Correct Keypoint (PCK) %\% for OpenMonkeyStudio dataset. Following [4], the threshold for considering a keypoint to be correct is set at 1010 cm.
Refer to caption
Figure 4: Qualitative 3D reconstruction comparison between the multi-view triangulation technique and our technique for Monkey body [4] and human hands [55] when operated over noisy 2D keypoints.

Evaluation analysis

For the Monkey body dataset, multi-view 3D reconstruction with 22- or 33- view using our approach significantly outperforms the given results in [4] and achieves comparable fidelity with limited physical views. We consider all the keypoints as correct if their reconstruction is within 10cm10\textit{cm} of the groundtruth in the PCK protocol. Table 3 and top-right plot in Fig. 1 shows that we outperform the given results of 2-Views by a significant margin (1.2%1.2\% vs. 68.63%68.63\%). The fidelity of 3D reconstructions using our method continues to rise as we add in more views - evident by the uptick in performance from 33-views. Qualitative performance of Monkey and Human hands dataset is shown in Fig. 4 and quantitative performance of Monkey body is given in Tab. 1 when operated over noisy 2D keypoints. For the Human body dataset, we inject noise in the camera extrinsics, intrinsics, and 2D keypoints separately and compare the performance in Fig. 5 and Tab. 2. The baseline method fails when noise with a small standard deviation is added, degrading the fidelity of the 3D reconstruction. Since our approach is only dependent on the quality of 2D keypoints, it shows slightly degraded performance when the noise is injected over the input 2D keypoints. Qualitative 3D reconstruction of our approach in Fig. 45 shows the visual improvement over the classical multi-view triangulation approaches when operated over noisy 2D keypoints.

Refer to caption
Figure 5: Qualitative results on Human 3.6M dataset with σ=[0.5,0.5,25]\sigma=[0.5,0.5,25] as intrinsics, extrinsics, and 2D keypoints Gaussian noise, respectively.

5.2 Generation to unseen 2D data

Several multi-view approaches exists for 3D human pose estimation that leverage either full or weak 3D supervision [18, 43, 21, 49, 41, 44, 25]. None of these references, however, directly tackle the unsupervised multi-view 3D reconstruction problem and hence are not as general as our solution. However, to showcase the generalization capability of our approach, we include these approaches in our evaluation, shown in Tab. 4 (the values are in mm). Furthermore, we also compare against recent monocular unsupervised 3D reconstruction methods. We leverage the processed datasets by Dovotny et al. [37] as the detected 2D keypoints for a fair evaluation and use the evaluation split of H3.6M dataset for this comparison. We find that our approach outperforms all other unsupervised approaches, and is on-par with many supervised methods.

Method Detected 2D GT 2D
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Iskakov et al. [18] 20.8 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Remelli et al. [43] 30.2 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Kadkhodamohammadi et al.  [21] 49.1 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Tome et al.  [49] 52.8 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Pavlakos et al.  [41] 56.9 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Multi-view Martinez  [35] 57.0 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Rhodin et al. [44] 51.6 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFCCC9Kocabas et al  [25] 45.04 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600Kocabas et al. (SS w/o R)  [25] 70.67 -
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600PRN  [40] 124.5 86.4
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600RepNet  [50] 65.1 38.2
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600Iqbal et al. [17] 69.1 -
\rowcolor[HTML]FFFFFF \rowcolor[HTML]FFFFFF \rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600Pose-GAN  [29] 173.2 130.9
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600Deep NRSfM  [27] - 104.2
\rowcolor[HTML]FFFFFF \cellcolor[HTML]FFF600C3DPO  [37] 153.0 95.6
\cellcolor[HTML]FFF600MV NRSfM (Ours) 45.2 30.2
Table 4: Generalization experiments. Red tint rows have 3D supervision. Yellow tint are unsupervised 3D reconstruction methods. Our method is on par with most 3D supervised methods, and outperforms all unsupervised methods.

6 Discussion and Conclusion

We propose a multi-view 2D-3D lifting architecture that incorporates neural shape prior using the recent advances of modern deep learning methods. Our contribution combines the ideas from multi-view geometry and recent monocular deep 3D lifting approaches – essentially leveraging the best features of both worlds. We also show the generalization capability of the proposed approach by generating accurate 3D reconstructions on unseen data. Although we require limited rigid views at any instant of time, our approach still requires multiple non-rigid atemporal views to enforce the proposed neural shape prior during training/optimization. Literature in domain of neural shape priors is extensive [37, 51] and new innovations are proposed constantly, and we believe we could leverage these innovations in our framework as part of future direction.

References

  • [1] Antonio Agudo, Melcior Pijoan, and Francesc Moreno-Noguer. Image collection pop-up: 3d reconstruction and clustering of rigid and non-rigid categories. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2607–2615, 2018.
  • [2] Ijaz Akhter, Yaser Sheikh, and Sohaib Khan. In defense of orthonormality constraints for nonrigid structure from motion. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 1534–1541. IEEE, 2009.
  • [3] Ijaz Akhter, Yaser Sheikh, Sohaib Khan, and Takeo Kanade. Nonrigid structure from motion in trajectory space. In Advances in neural information processing systems, pages 41–48, 2009.
  • [4] Praneet C Bala, Benjamin R Eisenreich, Seng Bum Michael Yoo, Benjamin Y Hayden, Hyun Soo Park, and Jan Zimmermann. Openmonkeystudio: automated markerless pose estimation in freely moving macaques. bioRxiv, 2020.
  • [5] Amir Beck and Marc Teboulle. A fast iterative shrinkage-thresholding algorithm with application to wavelet-based image deblurring. In 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 693–696. IEEE, 2009.
  • [6] Christoph Bregler, Aaron Hertzmann, and Henning Biermann. Recovering non-rigid 3d shape from image streams. In Proceedings IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No. PR00662), volume 2, pages 690–696. IEEE, 2000.
  • [7] Hristos S Courellis, Samuel U Nummela, Michael Metke, Geoffrey W Diehl, Robert Bussell, Gert Cauwenberghs, and Cory T Miller. Spatial encoding in primate hippocampus during free navigation. PLoS biology, 17(12):e3000546, 2019.
  • [8] Yuchao Dai, Hongdong Li, and Mingyi He. A simple prior-free method for non-rigid structure-from-motion factorization. International Journal of Computer Vision, 107(2):101–122, 2014.
  • [9] Katerina Fragkiadaki, Marta Salas, Pablo Arbelaez, and Jitendra Malik. Grouping-based low-rank trajectory completion and 3d reconstruction. In Advances in Neural Information Processing Systems, pages 55–63, 2014.
  • [10] Richard A Gibbs, Jeffrey Rogers, et al. Evolutionary and biomedical insights from the rhesus macaque genome. science, 316(5822):222–234, 2007.
  • [11] Ralph Gross, Iain Matthews, Jeffrey Cohn, Takeo Kanade, and Simon Baker. Multi-pie. Image and Vision Computing, 28(5):807–813, 2010.
  • [12] Darcy L Hannibal, Eliza Bliss-Moreau, Jessica Vandeleest, Brenda McCowan, and John Capitanio. Laboratory rhesus macaque social housing and social changes: implications for research. American Journal of Primatology, 79(1):e22528, 2017.
  • [13] Richard Hartley and Andrew Zisserman. Multiple view geometry in computer vision. Cambridge university press, 2003.
  • [14] Richard I Hartley and Peter Sturm. Triangulation. Computer vision and image understanding, 68(2):146–157, 1997.
  • [15] Brian Hrolenok, Tucker Balch, David Byrd, Rebecca Roberts, Chanho Kim, James M Rehg, Scott Gilliland, and Kim Wallen. Use of position tracking to infer social structure in rhesus macaques. In Proceedings of the Fifth International Conference on Animal-Computer Interaction, pages 1–5, 2018.
  • [16] Catalin Ionescu, Dragos Papava, Vlad Olaru, and Cristian Sminchisescu. Human3.6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE transactions on pattern analysis and machine intelligence, 36(7):1325–1339, 2013.
  • [17] Umar Iqbal, Pavlo Molchanov, and Jan Kautz. Weakly-supervised 3d human pose learning via multi-view images in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5243–5252, 2020.
  • [18] Karim Iskakov, Egor Burkov, Victor Lempitsky, and Yury Malkov. Learnable triangulation of human pose. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7718–7727, 2019.
  • [19] Sebastian Hoppe Nesgaard Jensen, Mads Emil Brix Doest, Henrik Aanæs, and Alessio Del Bue. A benchmark and evaluation of non-rigid structure from motion. International Journal of Computer Vision, 129(4):882–899, 2021.
  • [20] Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, and Yaser Sheikh. Panoptic studio: A massively multiview system for social motion capture. In Proceedings of the IEEE International Conference on Computer Vision, pages 3334–3342, 2015.
  • [21] Abdolrahim Kadkhodamohammadi and Nicolas Padoy. A generalizable approach for multi-view 3d human pose regression. Machine Vision and Applications, 32(1):1–14, 2021.
  • [22] Sinead Kearney, Wenbin Li, Martin Parsons, Kwang In Kim, and Darren Cosker. Rgbd-dog: Predicting canine pose from rgbd sensors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8336–8345, 2020.
  • [23] Matt J Kessler, John D Berard, and Richard G Rawlins. Effect of tetanus toxoid inoculation on mortality in the cayo santiago macaque population. American journal of primatology, 15(2):93–101, 1988.
  • [24] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [25] Muhammed Kocabas, Salih Karagoz, and Emre Akbas. Self-supervised learning of 3d human pose using multi-view geometry. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1077–1086, 2019.
  • [26] Chen Kong and Simon Lucey. Prior-less compressible structure from motion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4123–4131, 2016.
  • [27] Chen Kong and Simon Lucey. Deep non-rigid structure from motion. In Proceedings of the IEEE International Conference on Computer Vision, pages 1558–1567, 2019.
  • [28] Chen Kong, Rui Zhu, Hamed Kiani, and Simon Lucey. Structure from category: a generic and prior-less approach. In 2016 Fourth International Conference on 3D Vision (3DV), pages 296–304. IEEE, 2016.
  • [29] Yasunori Kudo, Keisuke Ogaki, Yusuke Matsui, and Yuri Odagiri. Unsupervised adversarial learning of 3d human pose from 2d joint locations. arXiv preprint arXiv:1803.08244, 2018.
  • [30] Suryansh Kumar. Non-rigid structure from motion: Prior-free factorization method revisited. In Winter Conference on Applications of Computer Vision (WACV 2020), 2020.
  • [31] Suryansh Kumar. Non-rigid structure from motion: Prior-free factorization method revisited. In The IEEE Winter Conference on Applications of Computer Vision, pages 51–60, 2020.
  • [32] Suryansh Kumar, Yuchao Dai, and Hongdong Li. Multi-body non-rigid structure-from-motion. In 2016 Fourth International Conference on 3D Vision (3DV), pages 148–156. IEEE, 2016.
  • [33] Suryansh Kumar, Yuchao Dai, and Hongdong Li. Superpixel soup: Monocular dense 3d reconstruction of a complex dynamic scene. IEEE transactions on pattern analysis and machine intelligence, 2019.
  • [34] Seong Hun Lee and Javier Civera. Closed-form optimal two-view triangulation based on angular errors. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2681–2689, 2019.
  • [35] Julieta Martinez, Rayat Hossain, Javier Romero, and James J Little. A simple yet effective baseline for 3d human pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, pages 2640–2649, 2017.
  • [36] Alejandro Newell, Kaiyu Yang, and Jia Deng. Stacked hourglass networks for human pose estimation. In European conference on computer vision, pages 483–499. Springer, 2016.
  • [37] David Novotny, Nikhila Ravi, Benjamin Graham, Natalia Neverova, and Andrea Vedaldi. C3dpo: Canonical 3d pose networks for non-rigid structure from motion. In Proceedings of the IEEE International Conference on Computer Vision, pages 7688–7697, 2019.
  • [38] Vardan Papyan, Yaniv Romano, and Michael Elad. Convolutional neural networks analyzed via convolutional sparse coding. The Journal of Machine Learning Research, 18(1):2887–2938, 2017.
  • [39] Hyun Soo Park, Takaaki Shiratori, Iain Matthews, and Yaser Sheikh. 3d reconstruction of a moving point from a series of 2d projections. In European conference on computer vision, pages 158–171. Springer, 2010.
  • [40] Sungheon Park, Minsik Lee, and Nojun Kwak. Procrustean regression networks: Learning 3d structure of non-rigid objects from 2d annotations. In European Conference on Computer Vision, pages 1–18. Springer, 2020.
  • [41] Georgios Pavlakos, Xiaowei Zhou, Konstantinos G Derpanis, and Kostas Daniilidis. Harvesting multiple views for marker-less 3d human pose annotations. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6988–6997, 2017.
  • [42] Marc Pollefeys, Reinhard Koch, and Luc Van Gool. Self-calibration and metric reconstruction inspite of varying and unknown intrinsic camera parameters. International Journal of Computer Vision, 32(1):7–25, 1999.
  • [43] Edoardo Remelli, Shangchen Han, Sina Honari, Pascal Fua, and Robert Wang. Lightweight multi-view 3d pose estimation through camera-disentangled representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6040–6049, 2020.
  • [44] Helge Rhodin, Jörg Spörri, Isinsu Katircioglu, Victor Constantin, Frédéric Meyer, Erich Müller, Mathieu Salzmann, and Pascal Fua. Learning monocular 3d human pose estimation from multi-view images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8437–8446, 2018.
  • [45] Terence Sim, Simon Baker, and Maan Bsat. The cmu pose, illumination, and expression (pie) database. In Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition, pages 53–58. IEEE, 2002.
  • [46] Carsten Steger. Algorithms for the orthographic-n-point problem. Journal of Mathematical Imaging and Vision, 60(2):246–266, 2018.
  • [47] Ke Sun, Bin Xiao, Dong Liu, and Jingdong Wang. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5693–5703, 2019.
  • [48] Xiao Sun, Bin Xiao, Fangyin Wei, Shuang Liang, and Yichen Wei. Integral human pose regression. In Proceedings of the European Conference on Computer Vision (ECCV), pages 529–545, 2018.
  • [49] Denis Tome, Matteo Toso, Lourdes Agapito, and Chris Russell. Rethinking pose in 3d: Multi-stage refinement and recovery for markerless motion capture. In 2018 international conference on 3D vision (3DV), pages 474–483. IEEE, 2018.
  • [50] Bastian Wandt and Bodo Rosenhahn. Repnet: Weakly supervised training of an adversarial reprojection network for 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7782–7791, 2019.
  • [51] Chaoyang Wang, Chen-Hsuan Lin, and Simon Lucey. Deep nrsfm++: Towards 3d reconstruction in the wild. arXiv preprint arXiv:2001.10090, 2020.
  • [52] Xiaowei Zhou, Menglong Zhu, Spyridon Leonardos, Konstantinos G Derpanis, and Kostas Daniilidis. Sparseness meets deepness: 3d human pose estimation from monocular video. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4966–4975, 2016.
  • [53] Yingying Zhu, Dong Huang, Fernando De La Torre, and Simon Lucey. Complex non-rigid motion 3d reconstruction by union of subspaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1542–1549, 2014.
  • [54] Yingying Zhu and Simon Lucey. Convolutional sparse coding for trajectory reconstruction. IEEE transactions on pattern analysis and machine intelligence, 37(3):529–540, 2013.
  • [55] Christian Zimmermann, Duygu Ceylan, Jimei Yang, Bryan Russell, Max Argus, and Thomas Brox. Freihand: A dataset for markerless capture of hand pose and shape from single rgb images. In Proceedings of the IEEE International Conference on Computer Vision, pages 813–822, 2019.