Invertible Neural Skinning
Abstract
Building animatable and editable models of clothed humans from raw 3D scans and poses is a challenging problem. Existing reposing methods suffer from the limited expressiveness of Linear Blend Skinning (LBS), require costly mesh extraction to generate each new pose, and typically do not preserve surface correspondences across different poses. In this work, we introduce Invertible Neural Skinning (INS) to address these shortcomings. To maintain correspondences, we propose a Pose-conditioned Invertible Network (PIN) architecture, which extends the LBS process by learning additional pose-varying deformations. Next, we combine PIN with a differentiable LBS module to build an expressive and end-to-end Invertible Neural Skinning (INS) pipeline. We demonstrate the strong performance of our method by outperforming the state-of-the-art reposing techniques on clothed humans and preserving surface correspondences, while being an order of magnitude faster. We also perform an ablation study, which shows the usefulness of our pose-conditioning formulation, and our qualitative results display that INS can rectify artefacts introduced by LBS well.
1 Introduction
Being able to create animatable representations of clothed humans beyond skinned meshes is essential for building realistic augmented or virtual reality experiences and improving simulators. Towards this goal, we consider the task of building animatable human representations from raw 3D scans and corresponding poses. Prior work in this area has seen a shift from building parametric models of humans [31, 6, 22], to more recent works learning implicit 3D neural representations [13, 50, 51, 55, 12, 56, 1] from data in canonical space. These canonical representations are animated to a new pose by a learning skinning weight field around them [11, 14, 52, 57, 37] and applying Linear Blend Skinning (LBS) to warp the surface, where the pose is defined by a bone skeleton underlying the 3D surface.

These prior works generally suffer from the limited expressivity of LBS when handling complex pose-varying deformations, such as those of loose clothes and body tissue (i.e. muscle bulges, skin wrinkles). In parametric models like SMPL [31], such deformations are represented by adding simple linear pose correctives (aka blend shapes), but these are restrictive and only work for unclothed humans. Implicit methods, to relieve this issue, learn their canonical representations conditioned on the deformed pose [11, 14]. However, this conditioning comes with two major drawbacks during reposing. Given the sequence of poses, a new mesh has to be extracted from scratch for each pose, which becomes a bottleneck when animating subjects at a high frame-rate or resolution. Also, as a consequence of this step, correspondences (topology preservation) between the surfaces of the same subject across different poses are lost.
Invertible Neural Networks (INN) [16, 26, 15] are bijective functions that can preserve exact correspondences between their input and output spaces, while learning complex non-linear transforms between them. This ability of INNs makes them a suitable candidate for reposing, and in this work, we leverage INNs to build an Invertible Neural Skinning (INS) pipeline. For this, we first build a Pose-conditioned Invertible Network, abbreviated as PIN111We avoid the abbreviation PINN to avoid confusion with Physics-inspired Neural Networks [48], to learn pose-conditioned deformations. Next, to create an end-to-end Invertible Neural Skinning (INS) pipeline, we place two PINs around a differentiable LBS module, and use a pose-free canonical representation. These PINs help capture the non-linear surface deformations of clothes across poses and alleviate the volume loss suffered from the LBS operation. Since our canonical representation remains pose-free, we perform the expensive mesh extraction exactly once, and repose the mesh by simply warping it with the learned LBS and an inverse pass through PINs.
We demonstrate the strong performance of INS by outperforming the previous state-of-the-art reposing method SNARF [11]. On clothed humans data, we find INS provides an absolute gain of roughly 1% when compared to SNARF with pose-conditioning, and roughly 6% compared to SNARF without pose-conditioning. We conduct experiments on much simpler minimally clothed human data and obtain competitive results. We also find INS to be an order of magnitude faster at reposing long sequences. We ablate our INS and demonstrate the effectiveness of our pose-conditioning formulation. Our results clearly show that the proposed INS can correct the LBS artefacts well.
2 Related Work
Representing Articulate Characters in 3D. Over the years, a significant amount of prior work for building parametric representations of the human body [31, 22, 45, 40, 30, 61, 2] or for specific parts such as hands and faces [39, 49] was developed. Beyond humans, recent work developed parametric animal models [5, 65, 4]. Encouraged by the rapid progress in implicit neural 3D representations [38, 36], a number of works explored building implicit human representations with and without clothing [58, 23, 60, 63, 59, 53, 57, 33, 64, 20]. Representing characters as implicit functions comes with a cost of time-consuming mesh extraction via Marching Cubes [32].
Animating 3D Representations with Poses. Parametric models usually define the correspondences between poses, represented as a set of bones, and mesh vertices through Linear Blend Skinning (LBS) weights. These weights provide a soft assignment of vertices to human bones. Thus for animation, these models simply transform the vertices using a linear combination of bone transformations. When the parametric model is not available, these weights need to be discovered. To this end, recent works adopt learning-based solutions for discovering LBS weights [11, 10, 47, 46, 29, 37, 14, 52]. They usually assume a shared canonical space and learn a canonical LBS weight field, which is used for deforming the body in the novel pose during inference. However, at training time, the character needs to be warped backward from deformed to canonical space, i.e. given deformed points, we need to obtain corresponding canonical points. Thus some works [52, 8, 62, 17, 56] learn LBS weights separately in deformed and canonical spaces, which could be used for establishing correspondences. These generally require cycle-consistency losses for regularization. Recently, SNARF [11] proposed to compute these correspondences by finding the solutions of the LBS equation using an iterative solver. We adopt a similar formulation, to discover the correspondences as well.
However, LBS is often insufficient to capture non-linear deformations of flowy clothes and body tissue (i.e. muscle bulges). To mitigate this problem, prior works [11, 14, 58] condition their canonical representations on the deformed pose. Such conditioning helps to alleviate the shortcomings of pure-LBS deformations, but this comes at a cost following two major limitations:
-
•
Slow Reposing. To generate a new animation given a sequence of poses across time, these methods extract a separate mesh from scratch for each pose. This becomes a bottleneck if we want to pose the character at a high frame-rate or resolution.
-
•
No Correspondences. As a consequence of the above step, two completely separate meshes get extracted at each pose with no correspondence between them.
In our work, we address both these limitations by extending pure-LBS formulation with additional Pose-conditioned Invertible Networks (PIN), while using a pose-free canonical representation.
Invertible Neural Networks for 3D Vision. INNs [15, 16, 26, 21, 41, 18, 24] were initially designed for tractable density estimation of high-dimensional and generative modeling, a.k.a. Normalizing Flows [27, 3]. Usually, INNs are built by chaining together multiple conditional Coupling Layers [15, 16], where a single coupling layer defines an invertible transformation between its input and output. The main idea behind Coupling Layers is that if we split the input into two parts and only modify the first part while conditioning this modification on the second, this should be trivially invertible. Another popular type of invertible transformations are Invertible Residual layers [24] with small conditioning numbers. They utilize fixed point iterations for finding an inverse. In our work, we mostly rely on Coupling Layers since they are faster, and we did not see any additional benefits from Residual layers. In the context of 3D vision, INNs were explored for learning primitives of 3D representations [43], doing 3D shape-completion tasks [28], and reconstructing dynamic scenes [9]. However, to the best of our knowledge, their usage has not been explored for animating 3D characters.
3 Method
Task Setup. The goal of our work is to learn a human 3D representation that allows the generation of novel poses beyond original training data (a.k.a. reposing). For each subject, we assume the availability of pairs consisting of bone poses and 3d meshes denoted as . Such data can be obtained from human scans, and the poses can be estimated by fitting a parametric SMPL-like body model to these scans. Given this data, we wish to learn a subject-specific implicit neural representation in a canonical space and a method to animate this representation.
Deformed and Canonical Spaces. We denote an input point in deformed space as and a point in the canonical space as . Since our input consists of a sequence of deformed (posed) meshes, we use the superscript to indicate the time-step of capture. As our canonical space is independent of the pose, it is shared across all the time steps; hence, is not time-indexed.
Deformed and Canonical Poses. We follow the SMPL model [31], which represents body pose as a set of bones in a kinematic tree. While reposing, as we only require the relative pose between canonical and deformed space at any given time , we represent this by , where represents a transformation of the bone in 3D space, i.e. with corresponding rotation and translation . We denote the total number of bones by .

Pose-free Canonical Occupancy. To represent a specific subject, we use an Occupancy Network [36] conditioned solely on the input point . The canonical surface is then represented implicitly as a level-set () of this occupancy network.
(1) |
To extract this canonical iso-surface as a mesh, we use the MISE [36] algorithm. This is different from previous works [11, 14] that use additional pose-conditioning in the canonical occupancy network.
Sampling Points. For both training and evaluation of INS, we sample 3D points in deformed space and get their ground-truth occupancy values of zero or one based on whether they lie outside or the mesh (scan). We put exact details on this sampling in Appendix A.1.
3.1 Differentiable Forward Blend Skinning
To animate our subject from their canonical to deformed pose we use Linear Blend Skinning (LBS), which involves deforming the canonical surface according to a convex combination of rigid bone transforms. Specifically, we use the differentiable LBS formulation from SNARF [11] and summarize it below.
Canonical Weight Field. We define a learnable weight field in canonical space parameterized by a neural network, . For a given point in canonical space, this weight field predicts the blend weights corresponding to each bone:
(2) |
To make weights () convex for LBS, they are constrained to be always non-negative and sum to 1 using softmax.
LBS. Given the above weight field and the relative body pose as bone transforms , we can forward warp any point of our canonical space to deformed space using Linear Blend Skinning as follows:
(3) |
where represents the corresponding point in deformed space where lands after LBS.


Searching Canonical Correspondences. While training on raw scans, we are only provided with points in deformed space. To find their possible correspondences in canonical space, we solve for the roots of Equation 3 using an iterative solver while keeping constant. Specifically, we use Broyden’s Method [7] to find a set of point correspondences for each deformed point by initializing the root-finding algorithm at different points in the canonical space.
(4) |
Differentiable Skinning. The above formulation is end-to-end differentiable as it is possible to compute the gradients of the weight field with respect to input point via implicit differentiation as shown in SNARF [11], Section 3.4. In this work, we also extend these derivations to compute the gradient of correspondences w.r.t. input points. For more details, please refer to Appendix D.
Why is LBS insufficient? The above differentiable formulation suffers from the same limitations of traditional LBS, such as being unable to represent the clothed surfaces, and introducing volume loss, as shown in Figure 6, b). This is especially problematic when learning from real-world data of clothed humans in various poses.
3.2 Pose-conditioned Invertible Network (PIN)
Invertible networks [16, 15, 26] are bijective functions composed of modular components called coupling layers, which preserve 1-1 correspondences between their input and output. In this section, we describe the construction of our proposed pose-conditioned coupling layer, which is chained together to construct a PIN.
2D Coupling Layer (Figure 3). A coupling layer operates by splitting its input into two parts using a fixed breaking pattern. After splitting, the first part of the input is transformed by applying a sequence of invertible operations, such as translation and rotation. The parameters for these operations can be produced by any arbitrary function that is jointly conditioned on the second part of the input and an external conditioning, such as pose in our case.
Formally, as we operate in 3D space, let the input point be defined as , and the input splits are and . Then the 2D coupling layer defines an invertible transformation as follows:
(5) |
where , and is a rotation matrix and translation vector produced by any arbitrary function that takes as input only the bone pose and the coordinate . The inverse of the coupling layer can be computed easily by:
(6) |
We describe computation of operation parameters and next.
Pose Embedding. We encode every bone transform in pose using a MLP which takes a 6D input of concatenated 3D translation and rotation (as Euler angles). To obtain pose embedding, we concatenate the outputs of each bone as follows:
(7) |
Space Embedding. We use SIREN [54], a learned and periodic positional encoding, to map the spatial coordinates denoting it as
(8) |
We find that this helps to better represent high-frequency surface details such as cloth wrinkles.
Space and Pose Aware Conditioning (Figure 2). We observe that when the relative pose between deformed and canonical spaces is zero (i.e. , all bone transforms have identity rotation and zero translation), the coupling layer should not introduce any space-varying (i.e. -conditioned) changes.
To enforce this, we take the Hadamard product of the space and pose embeddings, and subsequently concatenate them obtaining
(9) |
We visualize the construction of our space and pose aware conditioning in Figure 2.
Rotation and Translation Maps. Finally, to produce parameters for coupling operations, we use two MLPs and , which take as input the above conditioning vector:
(10) | |||
(11) |
Note that the output of only predicts the angle of rotation in radians (a single value). The axis of rotation passes through the origin of the split input space, i.e. space here. We convert into a rotation matrix .
1D Coupling Layer. Unlike the 2D coupling layer described above, we cannot use the rotation operator in 1D, and in this case, we only use translation. For layer with split pattern as and the coupling operation becomes:
(12) |
where produced in a similar fashion as a 2D coupling layer using a translation map with single scalar output instead of 2D translation. The space embedding of Equation 8, in the 1D case, takes both coordinates as input .
IoU Surface | IoU Bounding Box | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Subject | Clothing | AVG-LBS | FIRST-LBS | SNARF | SNARF-NC | INS (ours) | AVG-LBS | FIRST-LBS | SNARF | SNARF-NC | INS (ours) |
Average | 65.01% | 57.41% | 72.24% | 66.89% | 73.13% | 65.12% | 57.5% | 72.17% | 66.78% | 73.19% |
Pose-conditioned Invertible Network (PIN). Finally, we compose our PIN by chaining together multiple 1D and 2D pose-conditioned coupling layers as:
(13) |
where represents point in 3D space, represents a coupling layer, and represents pose. Inverting PIN is simply equivalent to sequentially inverting each coupling layer in the reverse order:
(14) |
Since the PIN is invertible by construction, it preserves exact correspondences between its input and output spaces:
(15) |
We visualize a single coupling layer of PINin Figure 3.
3.3 Invertible Neural Skinning
Overview (Figure 4). Our overall posing pipeline INS is comprised of three previously described components chained together:
-
•
: A Pose-conditioned Invertible Network (PIN) that operates after canonical space and before LBS.
-
•
: A Pose-conditioned Invertible Network (PIN) that operates before deformed space and after LBS.
-
•
Differentiable LBS network as described in Section 3.1 operating between above PINs.
Next, we discuss how we formulate an invertible mapping that preserves correspondences between deformed and canonical spaces.
Deformed to Canonical (Training). For any point in deformed space, we first process it using PIN , and obtain . Next, we use Broyden’s algorithm to get correspondences of in canonical space, let’s say . Finally we use a second PIN to map these points in the pose-independent canonical space.
(16) |
To obtain the most suitable canonical correspondence, we take the over all predicted canonical occupancies
(17) |
During training, we approximate the with a function in order to backpropagate gradients softly through all correspondences following SNARF.
Training Objective. In our dataset, we are given points in deformed space and corresponding ground truth occupancy values of zero or one. We map these deformed points to canonical space and apply binary cross-entropy loss to jointly train all components of the posing network according to
(18) |
Auxiliary Objectives. Following SNARF, we enforce a prior on the canonical pose by using two additional losses during the first epoch. First, we sample additional points on bones in canonical pose and encourage their occupancies to be one. Second, we encourage the skinning weight of bone joints to be equal. However, no ground truth skinning weights are required during these steps.
Canonical to Deformed (Inference). Once trained, we can animate characters using INS in any given novel pose in two simple steps. First, running mesh extraction on the canonical occupancy network. Second, reposing the mesh vertices via an inverse pass of our posing pipeline as follow:
(19) |
Fast Reposing. As our canonical occupancy network is independent of we only have to extract mesh exactly once. And reposing this mesh for a sequence of poses simply becomes equivalent to performing multiple inferences described in Equation 19.
4 Experiments
IoU Surface | IoU Bounding Box | |||||
---|---|---|---|---|---|---|
Subject | SNARF | SNARF-NC | INS (ours) | SNARF | SNARF-NC | INS (ours) |
Average | 90.01% | 85.22% | 88.59% | 97.21% | 95.72% | 96.35% |
4.1 Evaluation
Datasets. Training our method requires sampled points in the deformed space, along with corresponding occupancies and poses. Thus, we benchmark INS on two datasets CAPE [34], which features scanned humans in loose clothing, and DFAUST [35], containing only minimally clothed human scans.
CAPE [34] contains scans of 15 subjects (8 males and 7 females), wearing 8 different types of garments while performing a large number of actions. These actions were recorded using a high-resolution body scanner (3dMD LLC, Atlanta, GA), and the scans were registered using an SMPL model [31]. Similarly to SNARF [11], INS requires training a new model every subject-cloth pair, and exhaustively training on every combination quickly becomes expensive. To manage computational costs, we use a subset of 15 sequences. This subset covers all garment types at least once and most of the subjects, thus capturing variations in both — body shape and clothing.
DFAUST is a subset of the AMASS [35] dataset consisting of 10 subjects who are minimally clothed. Each subject is scanned similarly to CAPE while performing 10 different actions. As the subjects wear minimal clothing, much of their motions can be represented accurately by rigid body transformations. As a consequence of little subject clothing, we observe that DFAUST contains significantly fewer pose-specific deformations. Thus we note that DFAUST dataset is not well-posed to test the true capabilities of INS.
Data Splits. For a given subject in DFAUST or a subject-clothing pair in CAPE, we are provided with multiple temporal sequences, each containing a different action. We divide these sequences in 9:1 ratio into train and test sets. This split is similar to SNARF [11]. More details about training sequences and garment types are provided in Appendix B.
Metrics. Following SNARF [11], we report the mean Intersection-over-Union of points sampled near the mesh surface (IoU surface), and of points sampled uniformly in space (IoU bbox).

4.2 Baselines
SNARF-NC. We use SNARF [11] without pose-conditioning in the canonical occupancy network as our first and primary baseline. For this, we only remove the pose-conditioning used by SNARF such that canonical space becomes pose-independent, i.e. . We do not make any other changes. This setting is comparable to INS as it allows for fast posing and preserves correspondences across different poses.
SNARF. We also compare INS to the original SNARF [11] which uses a pose-conditioned occupancy network, i.e. . However, we point out that the above pose-conditioned occupancy comes at the sacrifice of fast posing, by requiring expensive mesh extraction for each new pose while not preserving correspondences across them. These disadvantages make the direct comparison between INS and SNARF based solely on their performance a little lopsided.
To obtain SNARF and SNARF-NC results, we use the official codebase 222https://github.com/xuchen-ethz/snarf released by the authors.
AVG-LBS. In addition to the above strong learned baselines, we provide results on two simpler baselines, which use the SMPL-fitted LBS weights to unpose the meshes (scans) using forward skinning. For this, we simply take an average of all the canonicalized training meshes to generate a final canonical mesh and deform it to any unseen given pose using Foward LBS and SMPL weights.
FIRST-LBS. This baseline is similar to the AVG-LBS baseline described above and uses SMPL-fitted weight for reposing. Instead of using an average across all training meshes, it only uses the first mesh, thus containing lesser pose-conditioned details.

4.3 Main Results
CAPE. We demonstrate the results of INS on clothed human data in Table 1. Given the challenging nature of modeling cloth deformations contained in this dataset, we find that INS surpasses SNARF-NC (without pose-conditioning) on average by +6.24% and +6.41% absolute percentage points in Surface IoU and Bounding Box IoU respectively. Moreover, INS also outperforms vanilla SNARF with pose-conditioning by +0.89% and +1.02% absolute percentage points in Surface IoU and Bounding Box IoU, respectively, while also enjoying the benefits of fast posing, and matched correspondences across various poses. We observe that the simple aggregation baseline of AVG-LBS performs quite closely with SNARF-NC with a performance drop of only 1.88% and 1.66% percentage points between them. However, AVG-LBS benefits from using a strong prior of parametric SMPL model and corresponding fitted weights.
DFAUST. We also report results on much simpler minimally clothed humans from the DFAUST dataset in Table 2. We find that INS outperforms SNARF-NC (without pose-conditioning) on average by +3.37% and +0.63% absolute percentage points in Surface IoU and Bounding Box IoU metrics, respectively. When compared to SNARF with pose conditioning, we find INS lags behind by -1.42% and -0.86% absolute percentage points in Surface IoU and Bounding Box IoU metrics, respectively. Given the minimal clothing and few pose-conditioned non-linear effects in DFAUST, we hypothesize that this performance drop can be attributed to SNARF overfitting easily to this benchmark. We believe this result also reflects the importance of testing on many realistic datasets such as CAPE.
Timing Study. In Figure 5, we compare the times taken by SNARF and INS to repose a clothed character across a sequence of 125 different poses. A single mesh extraction pass with MISE [36] operating on the cube of resolution , takes nearly seconds. While reposing, SNARF performs this operation for every given pose, whereas INS requires mesh extraction only once. Reposing the extracted mesh INS takes seconds for an inference pass, which is an order of magnitude faster than SNARF.
# | Ablation | IoU Surface () | IoU Bounding Box() |
---|---|---|---|
1 | INS(vanilla) | 72.83 | 72.69 |
2 | w/o Pose Mul. | ||
3 | w/o SIREN | ||
4 | w/o Rotation | ||
5 | w/o | ||
6 | w/o | ||
7 | w/o LBS |
4.4 Ablations
We perform numerous ablations of our INS setup, the results of which are summarized in Table 3.
Multiplying Pose and Space Embeddings is important. Reformulating the pose conditioning by simple concatenation, i.e. instead of multiply-then-concatenate, i.e. leads to a significant performance drop of 11% points in both metrics (Row 2).
contributes much more than . The invertible networks do not contribute equally to the performance. Removing the canonical space PIN leads to a sharp drop in IoU Surface by 4.94%, when compared to removing the deformed space INN which drops performance by only 0.17% points (Rows 5,6). We attribute this partly to the fact that editing an LBS-deformed mesh introduces the additional complexity of resolving part-wise correspondences, such as locating the new positions of joints and limbs.

Replacing SIREN positional embedding with MLP hurts. IoU Surface drops by 3.16% if the learned sinusoidal embeddings are replaced by simple MLP layers. This happens as fine surface details such as cloth wrinkles get blurred when using MLP, which is prevented by SIREN.
PINs without rotation perform slightly worse. Removing 2D operations from our PINs leads to a drop in IoU Surface by 0.92%. This happens because twisting deformations in the surface have to be represented by only displacements, which has previously been shown to be difficult to learn [42].
Simply using PINs without LBS performs worse. Entirely removing the differential LBS module and relying solely on PINs to capture the full articulate motion results in a huge drop of % on both metrics (Row 7).
4.5 Qualitative Analysis
INS can represent finer details compared to SNARF. In Figure 6, we demonstrate a subject in a challenging novel pose from the CAPE dataset. We find that SNARF (leftmost) is unable to capture fine details of cloth wrinkles, while also missing fingers as highlighted by the markers. Whereas SNARF-NC (second left) struggles with LBS artefacts such as volume loss by shrinking the arched back (highlighted), and displaying candy-wrapper effects. Finally, INS (second right) is able to capture much sharper local details around the body joints, such as around the waist and neck.
PINs can represent pose-varying deformations well. In Figure 8, we tease apart the edits made by solely PIN in the canonical space (displayed in the top row) given two unseen target poses (shown in the bottom row). As highlighted in the figure, we find that PIN learns to introduce pose-varying deformations such as raising cloth outlines around the neck and shoulder joint, introducing dress wrinkles at near extremities, and even adjusting limbs such as orienting feet.
Extreme poses from PosePrior. The most loose-fitted sequence in CAPE is of subject 00375 wearing a blazer and trouser, we show it animated in extreme poses from Pose Prior in first two columns of Figure 7. We also visualize other subjects (both clothed and naked) in extreme poses as well. We find that INS produces much more realistic cloth deformations compared to SNARF.

5 Conclusion
In this work, we presented an invertible, end-to-end differentiable, and trainable pipeline called Invertible Neural Skinning for reposing humans. For this, we built a Pose-conditioned Invertible Network (PIN) that can handle non-linear surface deformations of clothes and skin well, while also retaining correspondences across different poses. By placing two PINs around a differentiable LBS network and using a pose-free occupancy network, we created INS. We show that INS outperforms previous methods on clothed humans, while staying competitive on simpler and minimally clothed humans. Since reposing with our method requires the expensive mesh extraction exactly once, INS provides a speed-up of an order of magnitude compared to previous methods when animating long pose sequences.
Future Work. While training INS, the correspondence search performed by the differentiable LBS module often becomes a bottleneck, and future works can explore the possibility of eliminating this module completely — by learning its behavior from data. Furthermore, the occupancy network can be replaced with a neural representation that can handle texture and lightning and thus learn directly from 2D images and videos instead of raw scans.
# | Hyperparameters | Value | # | Hyperparameters | Value | |
---|---|---|---|---|---|---|
1 | No. of Parameters in INS | 1.80M | 2 | No. of Coupling Layers in / | 18 | |
3 | No. of Parameters in PIN / | 0.41M | 4 | No. of Parameters in Occupancy Network | 0.46M | |
5 | No. of Parameters in LBS Network | 53K | 6 | No. of Parameters in Bone Encoder | 0.46M | |
7 | Pose Embedding Dimension | 120 | 8 | Space Embedding Dimension | 120 | |
9 | Space and Pose Embedding | 240 | 10 | PIN input and output Dimension | 3 | |
11 | Number of epochs | 250 | 12 | Optimizer | Adam | |
13 | Batch size (DFAUST/CAPE) | 12/8 | 14 | Learning rate (INNs/Rest) | 1e-4/1e-3 | |
15 | Warm-up learning rate factor | 0.2 | 16 | Warm-up iterations | 2400 | |
17 | No. of points per batch | 60000 | 18 | Gradient clipping (L-2 Norm) | 4.0 |
Appendix A Implementation Details
A.1 Sampling Points
Following SNARF, we sample 200K points at every frame of the sequence. Half of these points (100K) are near the mesh (scan) surface, which are obtained by first sampling points on the mesh surface via Poisson disk sampling and followed by displacement with isotropic Gaussian noise (of ). Remaining half (100K) points are sampled uniformly within a bounding box scaled to 110% of the original bounding box.
A.2 Hyperparameters and Training Details
We trained all our models on a single Tesla V100 GPU for 250 epochs, which took nearly 40 hours on average. We used a learning rate of to train the PINs, while using a learning rate of for remaining modules. We used Adam [25] optimizer, with a linear warmup and no learning rate decay. PyTorch [44] is used for all the experiments. Please refer to Table 4 for full list.
A.3 Metrics
Given set of sampled points to be evaluated, we can represent the joint tuple of any point, its ground truth occupancy (which can be either 0 or 1), and predicted occupancy as ( respectively. Then Intersection over Union (IoU) can be computed as follows:
(20) |
To convert predicted probability to binary occupancy, we simply check if it is greater than 0.5. IoU Bounding Box operates with points sampled uniformly in the space, whereas Surface IoU operates with points sampled close to the body as described in Section A.1.
Appendix B Data
CAPE. CAPE originally contains 15 subjects, with each subject wearing 1-6 different types of clothing, and performing 3-74 different actions. On average, it contains nearly 249 frames for every subject-cloth pair. Due to high variance as well as high number of subject-cloth pairs, we use a subset of CAPE which contains 15 sequences of 13 subjects containing all 8 different types of clothings. A clothing in CAPE is denoted by a joint string of upper and lower body garment, for example, a subject wearing a blazer and pants is annotated as blazerlong, and so on.
Appendix C Invertible Neural Network
C.1 Initialization
We found that initializing the Pose-conditioned Invertible Networks (PINs) as identity modules stabilizes training, and allows the LBS network to train better. For this, we initialize the weights and biases of the last layer of the operation maps and (shown in Figure 2) as zeros.
C.2 Volume Preservation
Previous works in INNs [16, 26, 15] operating on high-dimensional ( 512-d) spaces constrained the Jacobian between input and output to an triangular matrix. This ensured that the Jacobian determinant, used for density modeling, was not expensive to compute. Determinant of a triangular matrix is simply multiplication of its diagonal. However, this prevented these works from using 2D operators such as rotation. We note that such a requirement is unnecessary in our setting, where Jacobian determinant is not needed. Additionally, using rotations also helps to preserve volume between the input and output spaces.
Next, we show that our PINs consisting of only rotations and translations are volume preserving. Note that to show a transform is volume preserving it is sufficient to show that the determinant of the Jacobian of this transform is one. From Equations 5 and 10, we can express the transform represented by a single 2D coupling layer as:
Then Jacobian of this transform becomes:
(21) |
And determinant of the above Jacobian is one, i.e. . Since, our PINs are composed of chaining together such coupling layers described in Equation 13, the overall determinant is also one. Hence volume is preserved within PINs.
Appendix D Gradients
Training INS requires calculating gradients of the Binary Cross Entropoy (BCE) loss (Equation 18), with respect to all the components. Let the weights of PINs , , the LBS network , and the Occupancy network be denoted with , , , and respectively. Backpropagating through the occupancy network and the PIN is straightforward:
(22) | |||
(23) |
where is the predicted occupancy. While the gradients for LBS network and second PIN are:
(24) | |||
(25) |
where is the root of the Equation 17, and is the input point. Pytorch’s automatic differentiation can handle the gradients in Equations 22 and 23. However, to obtain gradients w.r.t. implicit differentiation is required, similar to SNARF:
(26) |
And we can find gradients of with respect to as follows:
(27) |
Appendix E Miscellaneous Failed Experiments
In order to help with a future research in this direction we list several ideas that have been tried in our project, which however did not improve performance.
E.1 Invertible Residual Layers
Idea. Beyond the invertible space-splitting layers, we also experimented with using invertible residual layers [18, 24]. These layers operate by limiting the Lipschitz constant of the residual branch, which has be less than one in order to guarantee invertibility. Inversion of these layers can be achieved using a fixed-point iteration method, with convergence rate exponential in the number of iterations.
Outcome. We tried chaining residual layers with coupling layers alternately, and also placing them in the start and end of the invertible networks. However, these setups performed close or worse than without using residual layers, and were slower due to the expensive inversion pass.
E.2 Coupling Layers with Scales
Idea. Previous works CaDeX [28], and NeuralParts [43] utilized invertible networks with scale and translation operations, following the architecture proposed in RealNVP [16]. Such a transform can be represented as:
Outcome. When using these layers in our experiments we encountered the following difficulties:
-
•
Unstable Training. Floating point overflows occurred frequently during training due to the exponential scaling term. Even after carefully tuned gradient clipping and learning rate schedules, we encountered frequent experiment failures.
-
•
Squashing Effect. Since the scaling operator can lead to very large outputs from INN, generally a sigmoid squashing layer is used at the end to restrict the input to a fixed range that matches output distribution. However, due to this sigmoid layer, the INN can no longer be initialized as an Identity layer, even when all the rotations are identity and translations are zero. This leads to a squashing artefacts in outputs.
- •
E.3 Pose-conditioned 3D rotation and translation layers
Idea. We tried learning pose-conditioned global 3D rotation and translation layers. We implemented it similar to the coupling layers to predict rotation and translation parameters, but without any space conditioning.
Outcome. We did not find significant gains using this, and decided against using them in the final version as they had a big memory footprint. These layers often rotate the canonical space creating issues during mesh extraction.
Appendix F Visuals (video link)
We place all the qualitative results in this video, and discuss its contents below.
[Video Part-1] Pose-varying deformations in INS. In the first part of the video, we visualize the deformations introduced by PINs and under varying target poses (shown in top right). Deformations introduced by are shown in the top-middle part, whereas those introduced by are shown in the bottom-left part shaded in green. We demonstrate that INS is able to handle complex deformations of clothing across poses.
[Video Part-2] Baseline Comparison. In the second part of the video, we compare our method INS against all the five baselines discussed in Section 4.2 of the main paper. While both the LBS baselines, and SNARF-NC suffers from artifacts, we see that INS performs much better than other methods.
[Video Part-3] INS Ablations. In the third part of the video, we visualize results from various ablations reported in Section 4.4 of the main paper. Here, we find that removing SIREN leads to an overly smooth surface, and removing the LBS network makes it harder for the network to learn limb movements correctly.
[Video Part-4] Texture Propagation. As INS can preserve correspondences across poses, it becomes possible to propagate mesh attributes such as texture across various time frames. We conducted an experiment to test this, where we applied texture to the pose-independent canonical mesh. Next, we propagated this texture through the INS network. We show the results of this experiment in the fourth (and last) part of the video. We found that the applied texture deformed realistically like clothing, while being consistent across all frames, and was free of jittering.
To contrast and compare with the above experiment, we conducted similar texture propagation using SNARF. Since, SNARF decodes a separate mesh at each time-step, we color this mesh using the same scheme for coloring INS canonical mesh above. Propagating this texture through the LBS block, we find that it frequently leads to jittery artefacts as the texture overflows across semantically different parts. For example, the texture patch E4 applied to the blazer in a particular frame, overflows onto pants in another frame, and so on.
F.1 1D and 2D Coupling Layers

We visualize both 1D and 2D coupling layers together in Figure 9 for better understanding. In 1D case the space-pose aware conditioning gets conditioned on the 2D input, which helps to improve expressiveness. In the 2D case, we can make edits on an entire 2D plane conditioned on the 1D input. We find out that these blocks provide complementary benefits, thus we utilize both of them in the final architecture.
Appendix G Miscellaneous
G.1 Training speed of INS comparison.
Training the LBS module (Broyden’s method) takes nearly 0.70s, while both PIN modules combined take 0.09s per iteration. Accounting for loss computation and backprop, single iteration of SNARF takes 1.16s, while INS takes 1.34s. Overall there is a 13.4% slowdown using INS.
Appendix H Comparison with IMAvatar
We can model non-rigid deformations using MLP and leverage Broyden’s method in training, similar to IMAvatar [20].
Experiment: We conducted a prelimnary study, where we used a second MLP to model pose-conditioned vertex offsets before the LBS module similar to IMAvatar. We reused the clothed sequence from ablation study, and name this experiment as SNARF-MLP. We also added two tricks from INS to make this setup work. First, we zero initialized the last layer of offset MLP; second, we used the space-pose aware conditioning described in Equation 19 (Section 3.2).
# | Experiment | IoU Surf. | IoU BBox | Train Iter (sec) |
---|---|---|---|---|
1 | SNARF-MLP | 63.66 | 63.74 | 1.80 |
2 | w/ Pose Mul. | 68.62 | 68.64 | 1.80 |
3 | SNARF-NC | 66.10 | 66.11 | 1.16 |
4 | INS | 72.83 | 72.69 | 1.34 |
Quantative (Table 1): We found SNARF-MLP performs subpar compared to INS (Row 1 vs 4), while training much slowly due to multiple MLP runs (44% slowdown). Moreover, we find our pose-conditioning to boosts performance (Row 2), while not using zero init. leads to divergence.

Qualitative (Fig.1, RHS): Visually, we found SNARF-MLP to result in unnatural deformations (highlighted). We hypothesize, this could be due to the complex landscape of the resulting function which might make Broyden’s vulnerable to local minima. Thus, points close to each other, which ideally should have close solutions, converge to different local minima producing bending artefacts. In contrast, PINs optimized with SGD can avoid local minima [19].
IoU Surface | IoU Bounding Box | |||||
---|---|---|---|---|---|---|
Subject | SNARF | SNARF-NC | INS (ours) | SNARF | SNARF-NC | INS (ours) |
03375 | 70.16% | 66.1% | 71.13% | 62.57% | 70.24% | 71.02% |
50007 | 90.28% | 83.9% | 86.63% | 97.77% | 96.16% | 96.11% |
50022 | 92.19% | 88.09% | 92.58% | 98.05% | 96.68% | 98.12% |
50026 | 91.13% | 80.54% | 89.26% | 97.67% | 94.37% | 97.13% |
50004 | 89.6% | 85.4% | 88.96% | 97.48% | 96.3% | 97.21% |
50009 | 87.05% | 83.87% | 85.77% | 95.89% | 94.63% | 93.47% |
50021 | 89.76% | 87.26% | 90.46% | 96.79% | 95.58% | 96.95% |
50025 | 90.95% | 86.12% | 91.59% | 97.35% | 95.82% | 97.55% |
50027 | 89.54% | 86.9% | 85.91% | 96.74% | 95.79% | 93.75% |
50002 | 89.25% | 84.41% | 85.98% | 97.55% | 96.67% | 97.02% |
50020 | 90.31% | 85.69% | 88.74% | 96.85% | 95.15% | 96.23% |
Average | 90.01% | 85.22% | 88.59% | 97.21% | 95.72% | 96.35% |
Appendix I Subjectwise result breakdown
Expanding on the results from main draft, we also report per-subject breakdown on DFAUST (Table 6) and CAPE (Table 7) datasets.
IoU Surface | IoU Bounding Box | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Subject | Clothing | AVG-LBS | FIRST-LBS | SNARF | SNARF-NC | INS (ours) | AVG-LBS | FIRST-LBS | SNARF | SNARF-NC | INS (ours) |
50002 | longshort | 84.80% | 87.89% | 47.34% | 96.56% | 97.50% | 63.86% | 66.42% | 85.41% | 84.02% | 89.57% |
03375 | blazerlong | 62.63% | 53.91% | 70.16% | 66.1% | 72.83% | 62.57% | 54.05% | 70.24% | 66.11% | 72.69% |
00215 | poloshort | 64.91% | 57.27% | 73.1% | 65.68% | 72.22% | 64.92% | 57.55% | 73.1% | 65.8% | 72.35% |
00096 | shirtlong | 63.48% | 52.07% | 75.85% | 67.92% | 73.48% | 63.55% | 52.15% | 75.77% | 67.97% | 73.56% |
00096 | shirtshort | 59.05% | 54.67% | 75.32% | 63.59% | 74.89% | 59.12% | 54.7% | 75.16% | 63.61% | 74.96% |
00096 | jerseyshort | 62.0% | 56.63% | 73.28% | 64.28% | 74.61% | 62.02% | 56.45% | 73.19% | 63.98% | 74.37% |
00134 | longlong | 65.98% | 61.5% | 73.96% | 67.65% | 78.97% | 65.91% | 61.45% | 73.96% | 67.73% | 78.93% |
03223 | shortshort | 74.74% | 60.66% | 81.42% | 77.56% | 82.63% | 74.81% | 60.67% | 81.33% | 77.66% | 82.76% |
03331 | longshort | 70.3% | 65.2% | 77.13% | 75.12% | 77.12% | 70.5% | 65.52% | 77.03% | 74.5% | 77.22% |
00127 | shortlong | 73.84% | 65.7% | 72.48% | 70.0% | 73.2% | 74.09% | 65.87% | 72.31% | 69.9% | 73.33% |
02474 | longshort | 60.87% | 54.37% | 70.16% | 62.39% | 71.64% | 60.94% | 54.41% | 70.14% | 62.26% | 71.76% |
03284 | longshort | 64.68% | 58.5% | 67.04% | 65.76% | 68.79% | 64.68% | 58.21% | 66.87% | 65.44% | 68.77% |
00032 | longshort | 64.05% | 59.63% | 69.23% | 64.47% | 69.04% | 64.51% | 59.67% | 69.21% | 64.63% | 69.14% |
00122 | shortlong | 56.85% | 45.74% | 64.67% | 60.98% | 64.85% | 57.01% | 45.85% | 64.56% | 60.87% | 65.06% |
03394 | longlong | 62.56% | 52.11% | 69.43% | 66.16% | 71.55% | 62.6% | 52.25% | 69.27% | 66.1% | 71.36% |
00159 | longshort | 69.27% | 63.22% | 70.31% | 65.64% | 71.1% | 69.52% | 63.76% | 70.35% | 65.19% | 71.58% |
Average | 65.01% | 57.41% | 72.24% | 66.89% | 73.13% | 65.12% | 57.5% | 72.17% | 66.78% | 73.19% |
References
- [1] Thiemo Alldieck, Hongyi Xu, and Cristian Sminchisescu. imGHUM: Implicit generative models of 3d human shape and articulated pose. In Int. Conf. Comput. Vis., 2021.
- [2] Dragomir Anguelov, Praveen Srinivasan, Daphne Koller, Sebastian Thrun, Jim Rodgers, and James Davis. Scape: shape completion and animation of people. In ACM SIGGRAPH 2005 Papers, pages 408–416, 2005.
- [3] Lynton Ardizzone, Jakob Kruse, Sebastian Wirkert, Daniel Rahner, Eric W. Pellegrini, Ralf S. Klessen, Lena Maier-Hein, Carsten Rother, and Ullrich Köthe. Analyzing inverse problems with invertible neural networks, 2018.
- [4] Benjamin Biggs, Oliver Boyne, James Charles, Andrew Fitzgibbon, and Roberto Cipolla. Who left the dogs out? 3d animal reconstruction with expectation maximization in the loop. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm, editors, Computer Vision – ECCV 2020, 2020.
- [5] Benjamin Biggs, Thomas Roddick, Andrew Fitzgibbon, and Roberto Cipolla. Creatures great and smal: Recovering the shape and motion of animals from video. In Asian Conference on Computer Vision, pages 3–19. Springer, 2018.
- [6] Federica Bogo, Angjoo Kanazawa, Christoph Lassner, Peter Gehler, Javier Romero, and Michael J. Black. Keep it SMPL: Automatic estimation of 3D human pose and shape from a single image. In Eur. Conf. Comput. Vis., 2016.
- [7] Charles G Broyden. A class of methods for solving nonlinear simultaneous equations. Mathematics of computation, pages 19(92):577–593, 1965.
- [8] Andrei Burov, Matthias Nießner, and Justus Thies. Dynamic surface function networks for clothed human bodies. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 10754–10764, October 2021.
- [9] Hongrui Cai, Wanquan Feng, Xuetao Feng, Yan Wang, and Juyong Zhang. Neural surface reconstruction of dynamic scenes with monocular rgb-d camera. In Thirty-sixth Conference on Neural Information Processing Systems (NeurIPS), 2022.
- [10] Xu Chen, Tianjian Jiang, Jie Song, Jinlong Yang, Michael J. Black, Andreas Geiger, and Otmar Hilliges. gdna: Towards generative detailed neural avatars, 2022.
- [11] Xu Chen, Yufeng Zheng, Michael J Black, Otmar Hilliges, and Andreas Geiger. Snarf: Differentiable forward skinning for animating non-rigid neural implicit shapes. arXiv preprint arXiv:2104.03953, 2021.
- [12] Enric Corona, Albert Pumarola, Guillem Alenyà, Gerard Pons-Moll, and Francesc Moreno-Noguer. SMPLicit: Topology-aware generative model for clothed people. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), June 2021.
- [13] Boyang Deng, JP Lewis, Timothy Jeruzalski, Gerard Pons-Moll, Geoffrey Hinton, Mohammad Norouzi, and Andrea Tagliasacchi. Neural articulated shape approximation. In The European Conference on Computer Vision (ECCV). Springer, August 2020.
- [14] Boyang Deng, John P Lewis, Timothy Jeruzalski, Gerard Pons-Moll, Geoffrey Hinton, Mohammad Norouzi, and Andrea Tagliasacchi. Nasa neural articulated shape approximation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part VII 16, pages 612–628. Springer, 2020.
- [15] Laurent Dinh, David Krueger, and Yoshua Bengio. Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516, 2014.
- [16] Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. Density estimation using real nvp. arXiv preprint arXiv:1605.08803, 2016.
- [17] Nikita Drobyshev, Jenya Chelishev, Taras Khakhulin, Aleksei Ivakhnenko, Victor Lempitsky, and Egor Zakharov. Megaportraits: One-shot megapixel neural head avatars. In Proceedings of the 30th ACM International Conference on Multimedia, 2022.
- [18] Behrmann et al. Invertible residual networks. In ICML, 2019.
- [19] Keskar et al. On large-batch training for deep learning: Generalization gap and sharp minima, 2017.
- [20] Zheng et al. I M Avatar: Implicit morphable head avatars from videos. In CVPR, 2022.
- [21] Mathieu Germain, Karol Gregor, Iain Murray, and Hugo Larochelle. Made: Masked autoencoder for distribution estimation. In International Conference on Machine Learning, pages 881–889. PMLR, 2015.
- [22] Riza Alp Guler and Iasonas Kokkinos. Holopose: Holistic 3d human reconstruction in-the-wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
- [23] Zeng Huang, Yuanlu Xu, Christoph Lassner, Hao Li, and Tony Tung. ARCH: animatable reconstruction of clothed humans. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pages 3090–3099. IEEE, 2020.
- [24] Jörn-Henrik Jacobsen, Arnold W.M. Smeulders, and Edouard Oyallon. i-revnet: Deep invertible networks. In International Conference on Learning Representations, 2018.
- [25] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Int. Conf. Learn. Represent., 2015.
- [26] Diederik P Kingma and Prafulla Dhariwal. Glow: Generative flow with invertible 1x1 convolutions. arXiv preprint arXiv:1807.03039, 2018.
- [27] Ivan Kobyzev, Simon J.D. Prince, and Marcus A. Brubaker. Normalizing flows: An introduction and review of current methods. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
- [28] Jiahui Lei and Kostas Daniilidis. Cadex: Learning canonical deformation coordinate space for dynamic surface representation via neural homeomorphism. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
- [29] Ruilong Li, Julian Tanke, Minh Vo, Michael Zollhofer, Jurgen Gall, Angjoo Kanazawa, and Christoph Lassner. Tava: Template-free animatable volumetric actors, 2022.
- [30] Tianye Li, Timo Bolkart, Michael. J. Black, Hao Li, and Javier Romero. Learning a model of facial shape and expression from 4D scans. ACM Transactions on Graphics, (Proc. SIGGRAPH Asia), 36(6):194:1–194:17, 2017.
- [31] Matthew Loper, Naureen Mahmood, Javier Romero, Gerard Pons-Moll, and Michael J. Black. SMPL: A skinned multi-person linear model. ACM Trans. Graphics (Proc. SIGGRAPH Asia), 34(6):248:1–248:16, Oct. 2015.
- [32] William E. Lorensen and Harvey E. Cline. Marching cubes: A high resolution 3D surface construction algorithm. In Proceedings of the 14th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1987, Anaheim, California, USA, July 27-31, 1987, pages 163–169. ACM, 1987.
- [33] Qianli Ma, Jinlong Yang, Michael J. Black, and Siyu Tang. Neural point-based shape modeling of humans in challenging clothing. In 2022 International Conference on 3D Vision (3DV), September 2022.
- [34] Qianli Ma, Jinlong Yang, Anurag Ranjan, Sergi Pujades, Gerard Pons-Moll, Siyu Tang, and Michael J. Black. Learning to dress 3d people in generative clothing. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pages 6468–6477. IEEE, 2020.
- [35] Naureen Mahmood, Nima Ghorbani, Nikolaus F. Troje, Gerard Pons-Moll, and Michael J. Black. AMASS: Archive of motion capture as surface shapes. In Int. Conf. Comput. Vis., 2019.
- [36] Lars M. Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3D reconstruction in function space. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pages 4460–4470. Computer Vision Foundation / IEEE, 2019.
- [37] Marko Mihajlovic, Yan Zhang, Michael J. Black, and Siyu Tang. LEAP: Learning articulated occupancy of people. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), June 2021.
- [38] Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In European conference on computer vision, pages 405–421. Springer, 2020.
- [39] Gyeongsik Moon, Takaaki Shiratori, and Kyoung Mu Lee. Deephandmesh: A weakly-supervised deep encoder-decoder framework for high-fidelity hand mesh modeling. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part II, volume 12347 of Lecture Notes in Computer Science, pages 440–455. Springer, 2020.
- [40] Ahmed A. A. Osman, Timo Bolkart, and Michael J. Black. STAR: sparse trained articulated human body regressor. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part VI, volume 12351 of Lecture Notes in Computer Science, pages 598–613. Springer, 2020.
- [41] George Papamakarios, Theo Pavlakou, and Iain Murray. Masked autoregressive flow for density estimation. arXiv preprint arXiv:1705.07057, 2017.
- [42] Keunhong Park, Utkarsh Sinha, Jonathan T. Barron, Sofien Bouaziz, Dan B Goldman, Steven M. Seitz, and Ricardo Martin-Brualla. Nerfies: Deformable neural radiance fields, 2020.
- [43] Despoina Paschalidou, Angelos Katharopoulos, Andreas Geiger, and Sanja Fidler. Neural parts: Learning expressive 3d shape abstractions with invertible neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3204–3215, 2021.
- [44] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32, pages 8024–8035. Curran Associates, Inc., 2019.
- [45] Georgios Pavlakos, Vasileios Choutas, Nima Ghorbani, Timo Bolkart, Ahmed A. A. Osman, Dimitrios Tzionas, and Michael J. Black. Expressive body capture: 3d hands, face, and body from a single image. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019.
- [46] Sida Peng, Junting Dong, Qianqian Wang, Shangzhan Zhang, Qing Shuai, Hujun Bao, and Xiaowei Zhou. Animatable neural radiance fields for human body modeling. arXiv preprint arXiv:2105.02872, 2021.
- [47] Sida Peng, Yuanqing Zhang, Yinghao Xu, Qianqian Wang, Qing Shuai, Hujun Bao, and Xiaowei Zhou. Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9054–9063, 2021.
- [48] Chengping Rao, Hao Sun, and Yang Liu. Physics informed deep learning for computational elastodynamics without labeled data, 2020.
- [49] Javier Romero, Dimitrios Tzionas, and Michael J Black. Embodied hands: Modeling and capturing hands and bodies together. ACM Transactions on Graphics (ToG), 36(6):1–17, 2017.
- [50] Shunsuke Saito, Zeng Huang, Ryota Natsume, Shigeo Morishima, Hao Li, and Angjoo Kanazawa. PIFu: Pixel-aligned implicit function for high-resolution clothed human digitization. In 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019, pages 2304–2314. IEEE, 2019.
- [51] Shunsuke Saito, Tomas Simon, Jason M. Saragih, and Hanbyul Joo. PIFuHD: Multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pages 81–90. IEEE, 2020.
- [52] Shunsuke Saito, Jinlong Yang, Qianli Ma, and Michael J Black. Scanimate: Weakly supervised learning of skinned clothed avatar networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2886–2897, 2021.
- [53] Ruizhi Shao, Hongwen Zhang, He Zhang, Mingjia Chen, Yanpei Cao, Tao Yu, and Yebin Liu. Doublefield: Bridging the neural surface and radiance fields for high-fidelity human reconstruction and rendering. In CVPR, 2022.
- [54] Vincent Sitzmann, Julien N. P. Martel, Alexander W. Bergman, David B. Lindell, and Gordon Wetzstein. Implicit neural representations with periodic activation functions. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
- [55] Shih-Yang Su, Timur M. Bagautdinov, and Helge Rhodin. Danbo: Disentangled articulated neural body representations via graph neural networks. ArXiv, abs/2205.01666, 2022.
- [56] Shaofei Wang, Marko Mihajlovic, Qianli Ma, Andreas Geiger, and Siyu Tang. Metaavatar: Learning animatable clothed human models from few depth images. arXiv preprint arXiv:2106.11944, 2021.
- [57] Shaofei Wang, Katja Schwarz, Andreas Geiger, and Siyu Tang. Arah: Animatable volume rendering of articulated human sdfs. In European Conference on Computer Vision, 2022.
- [58] Chung-Yi Weng, Brian Curless, Pratul P. Srinivasan, Jonathan T. Barron, and Ira Kemelmacher-Shlizerman. Humannerf: Free-viewpoint rendering of moving people from monocular video, 2022.
- [59] Yuliang Xiu, Jinlong Yang, Dimitrios Tzionas, and Michael J. Black. ICON: Implicit Clothed humans Obtained from Normals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13296–13306, June 2022.
- [60] Hongyi Xu, Thiemo Alldieck, and Cristian Sminchisescu. H-nerf: Neural radiance fields for rendering and temporal reconstruction of humans in motion, 2021.
- [61] Hongyi Xu, Eduard Gabriel Bazavan, Andrei Zanfir, William T. Freeman, Rahul Sukthankar, and Cristian Sminchisescu. GHUM & GHUML: generative 3D human shape and articulated pose models. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pages 6183–6192. IEEE, 2020.
- [62] Gengshan Yang, Minh Vo, Natalia Neverova, Deva Ramanan, Andrea Vedaldi, and Hanbyul Joo. Banmo: Building animatable 3d neural models from many casual videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2863–2873, 2022.
- [63] Hao Zhao, Jinsong Zhang, Yu-Kun Lai, Zerong Zheng, Yingdi Xie, Yebin Liu, and Kun Li. High-fidelity human avatars from a single rgb camera. In CVPR, 2022.
- [64] Zerong Zheng, Han Huang, Tao Yu, Hongwen Zhang, Yandong Guo, and Yebin Liu. Structured local radiance fields for human avatar modeling, 2022.
- [65] Silvia Zuffi, Angjoo Kanazawa, Tanya Berger-Wolf, and Michael J. Black. Three-d safari: Learning to estimate zebra pose, shape, and texture from images ”in the wild”. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019.