# classagnostic_reconstruction_of_dynamic_objects_from_videos__a366ad77.pdf Class-agnostic Reconstruction of Dynamic Objects from Videos Zhongzheng Ren , Xiaoming Zhao , Alexander G. Schwing University of Illinois at Urbana-Champaign https://jason718.github.io/redo We introduce REDO, a class-agnostic framework to REconstruct the Dynamic Objects from RGBD or calibrated videos. Compared to prior work, our problem setting is more realistic yet more challenging for three reasons: 1) due to occlusion or camera settings an object of interest may never be entirely visible, but we aim to reconstruct the complete shape; 2) we aim to handle different object dynamics including rigid motion, non-rigid motion, and articulation; 3) we aim to reconstruct different categories of objects with one unified framework. To address these challenges, we develop two novel modules. First, we introduce a canonical 4D implicit function which is pixel-aligned with aggregated temporal visual cues. Second, we develop a 4D transformation module which captures object dynamics to support temporal propagation and aggregation. We study the efficacy of REDO in extensive experiments on synthetic RGBD video datasets SAIL-VOS 3D and Deforming Things4D++, and on real-world video data 3DPW. We find REDO outperforms state-of-the-art dynamic reconstruction methods by a margin. In ablation studies we validate each developed component. 1 Introduction 4D (3D space + time) reconstruction of both the geometry and dynamics of different objects is a long-standing research problem, and is crucial for numerous applications across domains from robotics to augmented/virtual reality (AR/VR). However, complete and accurate 4D reconstruction from videos remains a great challenge for mainly three reasons: 1) partial visibility of objects due to occlusion or camera settings (e.g., out-of-view parts, non-observable surfaces); 2) complexity of the dynamics including rigid-motion (e.g., translation and rotation), non-rigid motion (deformation caused by external forces), and articulation; and 3) variability within and across object categories. Existing work addresses the above challenges by assuming complete visibility through a multi-view setting [33, 4], or by recovering only the observable surface rather than the complete shape of an object [57], or by ignoring rigid object motion and recovering only the articulation [59], or by building shape templates or priors specific to a particular object category like humans [47]. However, these assumptions also limit applicability of models to unconstrained videos in the wild, where these challenges are either infeasible or only met when taking special care during a video capture. In contrast, we aim to study the more challenging unconstrained 4D reconstruction setting where objects may never be entirely visible. Specifically, we deal with visual inputs that suffer from: 1) occlusion: the moving occluder and self-articulation cause occlusion to change across time; 2) cropped view: the camera view is limited and often fails to capture the complete and consistent appearance across time; 3) front-view only: due to limited camera motion, the back side of the objects are often not captured at all in the entire video. Moreover, we focus on different dynamic object-types Indicates equal contribution 35th Conference on Neural Information Processing Systems (Neur IPS 2021). Person (synthetic) Articulation, rigid motion Car/Motorcycle (synthetic) Non-rigid motion, rigid motion Person (real-world) Articulation, rigid motion Animal (synthetic) Articulation, rigid motion Figure 1: We present REDO, a 4D reconstruction framework that predicts the geometry and dynamics of various objects for a given video clip. Despite objects being either occluded (e.g., car/motorcycle) or partially-observed, REDO recovers relatively complete and temporally smooth results. with complex motion patterns. These objects could either move in 3D space, or be deformed due to external forces, or articulate themselves. Importantly, we aim for a class-agnostic reconstruction framework which can recover the accurate shape at each time-step. To achieve this we develop REDO. As illustrated in Fig. 1, REDO predicts the shape of different objects (e.g., human, animals, car) and models their dynamics (e.g., articulation, non-rigid motion, rigid motion) given input video clips. Besides the RGB frames, REDO takes as input the depth map, masks of the objects of interest, and camera matrices. In practice these inputs are realistic as depth-sensors are increasingly prevalent [19, 78] and as segmentation models [44, 17] are increasingly accurate and readily available, e.g., on mobile devices. If this data isn t accessible, off-the-shelf tools are applicable (e.g., Sf M [71], instance segmentation [26], video depth estimation [51]). To address the partial visibility challenge introduced by occlusion or camera settings, REDO predicts a temporally coherent appearance in a canonical space. To ensure that the same model is able to reconstruct very different object types in a unified manner we introduce a pixel-aligned 4D implicit representation, which encourages the predicted shapes to closely align with the 2D visual inputs ( 3.1). The visible parts from different frames of the video clip are aggregated to reconstruct the same object ( 3.2). During inference, the reconstructed object in canonical space is propagated to other frames to ensure a temporally coherent prediction ( 3.3). REDO achieves state-of-the-art results on various benchmarks ( 4). We first conduct experiments on two synthetic RGBD video datasets: SAIL-VOS 3D [27] and Deforming Things4D++ [42]. REDO improves over prior 4D reconstruction work [59, 70] by a great margin (+5.9 m Io U, - 0.085 m Cham., -0.22 m ACD on SAIL-VOS 3D and +2.2 m Io U, -0.063 m Cham., -0.047 m ACD on Deforming Things4D++ over OFlow [59]). We then test on the real-world calibrated video dataset 3DPW [86]. We find that REDO generalizes well and consistently outperforms prior 4D reconstruction methods (+10.1 m Io U, -0.124 m Cham., -0.061 m ACD over OFlow). We provide a comprehensive analysis to validate the effectiveness of each of the introduced components. 2 Related work In this section, we first discuss possible geometry representations. We then review fusion-based and learning-based 4D reconstruction approaches, followed by a brief introduction of Motion Capture methods. Lastly, we discuss works of dynamics modelling and related 4D reconstruction datasets. Geometric representations. Representations to describe 3D objects can be categorized into two groups: discrete and continuous. Common discrete representations are voxel grids [22, 90, 8], octrees [68, 82], volumetric signed distance functions (SDFs) [10, 30, 56], point-clouds [1, 18, 67], and meshes [23, 34, 55, 38, 88]. Even though being widely used, these representations pose important challenges. Voxel grids and volumetric-SDFs can be easily processed with deep learning frameworks, but are memory inefficient [76, 52, 64]. Point-clouds are more memory efficient to process [65, 66], but do not contain any surface information and thus fail to capture fine details. Meshes are more expressive, but their topology and discretization introduce additional challenges. To overcome these issues, continuous representations, i.e., parametric implicit functions, are introduced to describe the 3D geometry of objects [7, 60, 53] and scenes [54, 73]. These methods are not constrained by discretization and can thus model arbitrary geometry at high resolution [70, 81]. In Transformer $%&& ( 3.2)) Flow field Φ ( 3.2+) Flow field Φ ( 3.2+) Canonical space ["", "!, "#] Time embedding Transformer Dynamic feat 4' Static feat 5' Figure 2: Framework overview. For a query point p in canonical space, REDO first computes the pixel-aligned features from the feature maps of different frames using the flow-field Φ. It then aggregates these features using the temporal aggregator fagg. The obtained dynamic feature xp is eventually used to compute the occupancy score for shape reconstruction. practice, discretized representations like a mesh can be easily extracted from implicit functions via algorithms like Marching Cubes [49]. Fusion-based 4D reconstruction. Dynamic Fusion [57] is the pioneering work for reconstructing non-rigid scenes from depth videos in real-time. It fuses multiple depth frames into a canonical frame to recover the observable surface and adopt a dense volumetric 6D motion field for modeling motion. Another early work [14] reconstructs dynamic objects without explicit motion fields. Improvements of Dynamic Fusion leverage more RGB information [28], introduce extra regularization [74, 75], develop efficient representations [20], and predict additional surface properties [24]. Importantly, different from the proposed direction, these works only recover the observable surfaces rather then the complete shape. Moreover, these works often fail to handle fast motion and changing occlusion. Note, these models often have no trainable parameters and are more geometry-based. Learning-based 4D reconstruction. Reconstruction of the complete shape of dynamic objects is considered hard due to big intra-class variations. For popular object categories such as humans or certain animals, supervised learning of object template parameters has been utilized to ease reconstruction [36, 94, 47, 96, 97]. These templates are carefully designed parametric 3D models, which restrict learning to a low-dimensional solution space. However, the expressiveness of templates is limited. For instance, the SMPL [47] human template struggles to capture the clothing or hair styles. In addition, systems relying on templates are class-specific as different object categories need to be parameterized differently. We think it remains illusive to construct a template for every object category. To overcome these issues, OFlow [59] directly learns 4D reconstruction from scratch. However, it predicts articulation in normalized model space and thus overlooks rigid-motion of the object. It also struggles to handle occlusion. When ground-truth 3D models are not available, dynamic neural radiance fields (Ne RF) [54, 43, 63, 61] learn an implicit scene representation from videos, but are often scene-specific and don t scale to a class-agnostic setting. Self-supervised methods [91, 35, 41] are promising and learn 4D reconstruction via 2D supervision of differentiable rendering [37, 48]. In contrast, in this work, we present a class-agnostic and template-free framework which learns to recover the shape and dynamics from input videos. Motion/Performance capture. When restricting ourselves to human modeling, 4D reconstruction is often referred to as motion capture (Mo Cap). Marker and scanner based Mo Cap systems [62, 46, 40] are popular in industry, yet inconvenient and expensive. Without expensive sensors, camera-based methods are more applicable but often less competitive. Standard methods rely on a multi-view camera setup and leverage 3D human body models [32, 95, 21, 85]. Active research problems in the Mo Cap field include template-free methods [16, 15, 11], single-view depth camera methods [2, 92, 93], monocular RGB video methods [25, 79], and fine-grained body parts (e.g., hand, hair, face) methods [72, 50, 89, 33]. In contrast, this paper studies a more generic class-agnostic setting where multiple different non-rigid objects are reconstructed using one unified framework. Dynamics modeling. In different communities, the task of dynamics modeling is named differently. For 2D dynamics in images, optical flow methods [13, 80, 83] have been widely studied and used. Scene flow methods [43, 45, 31] aim to capture the dynamics of all points in 3D space densely, and are often time-consuming and inefficient. For objects, non-rigid tracking or matching methods [4, 77, 9, 5] study the dynamics of non-rigid objects. However, these methods often only track the observable surface rather then the complete shape space. Recently, neural implicit functions [43, 59] have been applied to estimate 3D dynamics, which we adapt and further improve through conditioning on pixel-aligned temporally-aggregated visual features. 4D reconstruction datasets. To support 4D reconstruction, a dataset needs to have both groundtruth 3D models and accurate temporal correspondences. Collecting such a dataset in a real-world setting is extremely challenging as it requires either expensive sensors or restricted experimental settings. To simplify the setting, existing data is either class-specific (e.g., human) [29, 87, 3], or of extremely small scale [74, 28], or lacking ground-truth 3D models [5]. Given the progress in computer graphics, synthetic data is becoming increasingly photo-realistic and readily available. Deforming Things4D [42] provides a large collections of humanoid and animal 4D models, but lacks textures and background. SAIL-VOS 3D [27] contains photo-realistic game videos together with various ground-truth. For more details, please see the comparison table in Appendix B. We aim to recover the 3D geometry of a dynamic object over space and time given a RGBD video together with instance masks and camera matrices. To achieve this we develop REDO which is illustrated in Fig. 2. Specifically, REDO reconstructs an object shape and its dynamics in a canonical space, which we detail in 3.1. For this REDO employs a temporal feature aggregator, a pixel-aligned visual feature extractor, and a flow-field, all of which we discuss in 3.2. These three modules help align dynamics of the object across time and resolve occlusions to condense the most useful information into the canonical space. Lastly, we detail inference ( 3.3) and training ( 3.4). Notation. The input of REDO is a fixed-length video clip denoted via {I1, . . . , IN}. It consists of N RGBD frames Ii 2 [0, 1]4 W H (i 2 {1, . . . , N}) of width W and height H, each recorded at timestep ti. For each frame Ii, we also assume the camera matrix and instance mask mij 2 {0, 1}W H are given, where mij indicates the set of pixels that correspond to object j. For readability, we define the following operations: 1) propagate: transforms temporally in 3D; 2) project: transforms from 3D to 2D (image space); and 3) lift: transforms from 2D (image space) to 3D. 3.1 Canonical 4D implicit function 2D points lifting Figure 3: Canonical Space. Depth points from different frames are lifted and aggregated to form an enlarged space for capturing the dynamic objects. Dynamic objects deform and move across time and are often occluded or partially out-of-view. In addition, the depth-size ambiguity also challenges reconstruction algorithms. To condense information about the object across both space and time, we leverage a canonical space C R3, which aims to capture a holistic geometry representation centered around the object of interest. In our case C denotes a volume-constrained 3dimensional space. Temporally, the canonical space corresponds to the center frame Ic at timestep tc with c = d(1 + N)/2e. To construct the canonical space, we infer a 3D volume around the dynamic object j. For all frames Ii, i 2 {1, . . . , N}, we first lift the pixels of instance mask mij of object j to the world space using the depth map and the camera matrices. These 3D points are then aggregated into canonical space using the camera matrix of the canonical frame. From the aggregated point clouds we infer the horizontal and vertical bounding box coordinates as well as the object s closest distance from the camera, i.e., Znear. However, since the visible pixels only represent the front surface of the observed object, the largest distance of the object from the camera, i.e., Zfar, is unknown. To estimate this value, we simply set Zfar to stay a fixed distance from Znear. We illustrate an example of this canonical space in Fig. 3, where the inferred canonical space (blue volume) is relatively tight and big enough to capture the complete shape and dynamics of the observed object. To more precisely capture an object inside the volume-constrained canonical space, we use a 4D implicit function. For each 3D query point in canonical space, this function conditions on a temporal feature to indicate whether that point is inside or outside the object. Specifically, for a query point inside the canonical space, i.e., for p 2 C, the 4D implicit function is defined as: g (p, xp) : C RK ! [0, 1]. (1) Intuitively, the higher the score g ( , ), the more likely the point p is part of the object. Here, subsumes all trainable parameters in the framework and xp 2 RK is a K-dimensional dynamic feature which summarizes information from all input frames. Concretely, {fenc( (Φ(p, tc, ti, v )) , Ii)}N where ( ) : C ! R3 transforms a point in canonical space to world space. Meanwhile, Φ is the flow-field which propagates the point p from tc to ti in canonical space, leveraging a learned 3-dimensional velocity-field v . Note, tc refers to the time-step of the canonical frame Ic while ti denotes the time-step of frame Ii. Given the transformed point (Φ(p, tc, ti, v )) 2 R3, we use a pixel-aligned feature encoder fenc to extract a set of visual representations, each of which comes from one of the N frames Ii, i.e., 8i 2 {1, . . . , N}. The set function fagg is a temporal feature aggregator which merges the temporal information of different time-steps. We ll discuss details of temporal aggregator fagg, encoder fenc, flow-field Φ and velocity-field v in 3.2. This canonical 4D implicit function propagates points across time and extracts the pixel-aligned visual representations from different frames. This differs from: 1) prior works that only consider static objects. Actually, setting N = 1 in Eq. (2) will simplify Eq. (1) to g(p, fenc( (p), I)), recovering a static Pixel-aligned Implicit Function (PIFu) used in [69, 70]; 2) OFlow [59] which encodes the whole video clip {I1, . . . , IN} into one single feature vector, thus loosing spatial information. 3.2 Framework design In this section, we introduce the temporal aggregator fagg, the feature extractor fenc, and the flow-field Φ used in Eq. (2). These modules help REDO aggregate information from different time-steps (frames) and model the complex dynamics. a) Temporal aggregator fagg. To deal with partial visibility caused by occlusions or camera settings, we develop a transformer-based temporal aggregator fagg as shown in Fig. 2. fagg is a set function which computes K-dimensional point features xp 2 RK. Assume the flow-field Φ is given and object j is of interest. We first propagate a query point within the canonical space, i.e., the point p 2 C, to every time-step ti, obtaining locations (Φ(p, tc, ti, v )) 2 R3 8i 2 {1, . . . , N}. We then project each 3D point (Φ(p, tc, ti, v )) back to the corresponding 2D image frame Ii using the associated camera matrix. For all points that are projected into the mask area mij, we extract a visual representation using the pixel-aligned feature extractor fenc, which we detail below. Note that due to partial visibility, we may not be able to extract features from every frame in the clip. To cope with this, the aggregator fagg is designed as a transformer-based [84] set function. For more information, please see the implementation details in 4.1 and Appendix A. b) Pixel-aligned feature extractor fenc. Pixel-aligned features help REDO make 3D predictions that are aligned with the visual 2D input. To achieve this, we develop fenc(q, Ii) where the first argument is a point q in 3D world space and the second argument is a frame Ii. We first project the point q in world space to the frame Ii, and then use a pre-trained convolutional neural net [58] to extract the 2D feature map of a video frame Ii. For points that fall within the instance mask of a frame, we extract its visual representation using bi-linear interpolation at the projected location. We then append a positional encoding [84] of the frame time-step ti to this feature which helps to retain temporal information. We illustrate this process in Fig. 2. The resulting feature is combined with visual cues from other frames to serve as input to the temporal aggregator fagg. c) Flow-field Φ. The flow-field models object dynamics in space and time. For this let Φ(p, t1, t2, v ) 2 C denote a flow-field function in canonical space. It computes the position at time-step t2 of a 3D point, whose location is p 2 C at time-step t1. To compute the displacement, we define a velocity field v ( ) which represents the 3D velocity vectors in space and time via v (p, zp, t) : C RK R ! C. (3) Here, p 2 C is a point in canonical space with corresponding static feature zp 2 RK computed as {fenc( (p), Ii)}N Here, fenc( (p), Ii) is the feature of world coordinate point (p) extracted from frame Ii. Note, zp differs from xp defined in Eq. (2). The feature zp summarizes information of static locations from all frames Ii. This feature is beneficial as it helps capture whether a point remains static or whether it moves. The velocity network then leverages this feature zp to predict object dynamics. Using the velocity field, we compute the target location at time-step t2 of a point originating from location p at time-step t1 by integrating the velocity field over the interval [t1, t2] via Φ(p, t1, t2, v ) = p + v (Φ(p, t1, t, v ), zp, t)dt. (5) Note that Φ(p, t1, t2, v ) can represent both forward (t2 > t1) or backward motion (t2 < t1) given the initial location and velocity-field. To solve the flow-field for discrete video time-steps, we approximate the above continuous integral equation using a neural-ODE solver [6]. 3.3 Inference To reconstruct objects densely in a clip, the reconstruction/inference procedure is summarized as: Step 1: We construct the canonical space at the center frame Ic and sample a query point set P C uniformly from the inferred space. This step requires no network inference and is very efficient. Step 2: We extract static features zp for all p 2 P using Eq. (4). Next, we compute the trajectory of point p, i.e., Φ(p, tc, ti, v ) for all time-steps ti associated with frames Ii, 8i 2 {1, . . . , N}, using a neural ODE solver which solves Eq. (5). Step 3: We then compute dynamic features xp for all p 2 P using Eq. (2). Step 4: Using Eq. (1) we finally compute the occupancy scores for all p 2 P, which are then transformed into a triangle mesh via Multi-resolution Iso-Surface Extraction (MISE) [53]. Note, the mesh is constructed in canonical space. Step 5: To obtain the mesh associated with frame Ii, following Step 2, we use the flow-field Φ to propagate all vertices of the extracted mesh from the time-step tc of the canonical space to time-step ti which corresponds to non-canonical frame Ii. After using the function ( ) defined in Eq. (2), we obtain the mesh in 3D world space. While various discrete representations could be obtained from our implicit representation, we use meshes for evaluation and visualization purposes. 3.4 Training REDO is fully differentiable containing the following parametric components: the temporal aggregator fagg, the feature extractor fenc, the velocity-field network v , and the reconstruction network g . For simplicity, we use to subsume all trainable parameters of REDO. To better extract shape and dynamics from given video clips, we train REDO end-to-end using Lshape(D, ) + Ltemp(D, ), (6) where D is the training set. Lshape(D, ) is the shape reconstruction loss in canonical space which encourages REDO to recover the accurate 3D geometry. Ltemp(D, ) is a temporal coherence loss defined on temporal point correspondences. This loss encourages the flow-field Φ to capture the precise dynamics of objects. We detail training set, sampling procedure, and both losses next. Training set. We train REDO on a dataset D that consists of entries from different videos and of various object instances. Formally, D = {(Ii, ti)}N . Specifically, a data entry for an N-frame video clip includes three components: 1) a set of RGBD frames {Ii}, associated time-steps {ti}, as well as corresponding camera matrices 8i 2 {1, . . . , N}; 2) an instance ground-truth mesh V that provides-temporally aligned supervision. For every vertex v 2 V, we use vi to denote its position at time-step ti ; 3) the ground-truth occupancy Y in the canonical space. Concretely, for a point p 2 C in canonical space, occupancy label y(p) 2 {0, 1} indicates whether p is inside the object (y = 1) or outside the object (y = 0). As mentioned in 3.1, the canonical space is chosen so that it corresponds to time-step tc, where c = d(1 + N)/2e. Sampling procedure. To optimize for the parameters during training we randomly sample a set of points p 2 P(V) within the canonical space. P(V) contains a mixture of uniform sampling and importance sampling around the ground-truth mesh s surface V at time-step tc. Similar strategies are also used in prior works [69, 70]. Shape reconstruction loss. To encourage that the canonical 4D implicit function g accurately captures the shape of objects we use the shape reconstruction loss Lshape(D, ) = ({(Ii,ti)}N BCE (g (p, xp), y(p)) , (7) where BCE( , ) represents the standard binary cross-entropy loss. Temporal coherence loss. REDO models dynamics of objects explicitly through the flow-field Φ, which leverages the velocity-field network v . As the ground-truth correspondences across time are available in V, we define the temporal correspondence loss via the squared error Ltemp(D, ) = ({(Ii,ti)}N kΦ(vc, tc, ti, v ) vik2 4 Experiments We first introduce the key implementation details ( 4.1) and the experimental setup ( 4.2), followed by the quantitative results ( 4.3), qualitative results ( 4.5), and an in-depth analysis ( 4.4). 4.1 Implementation details We briefly introduce the key implementation details. Check Appendix A for a more detailed version. Input: We assume all input clips are trimmed to have N = 17 frames following [59]. During training, clips are randomly sampled from original videos which have any length. For videos that are shorter than 17 frames, we pad at both ends with duplicated starting and ending frames to form the clips. The validation and test set consist of fixed 17-frame clips. This simplified input setting allows us to split development into manageable pieces, and allows fair comparison with prior work. In practice, dense reconstruction on the entire video is achieved via a sliding window method. Reconstruction network: Following [69], the reconstruction network g is implemented as a 6-layer MLP with dimensions (259, 1024, 512, 256, 128, 1) and skip-connections. The first layer s dimension of 259 is due to the concatenation of visual features (256-dim) and query point locations (3-dim). Temporal aggregator: fagg uses a transformer model with 3 multi-headed self-attention blocks and a 1-layer MLP. Group normalization and skip-connections are applied for each block, and we set the hidden dimension to be 128. To compute the time-encoding, we use positional-encoding [84] with 6 exponentially increasing frequencies. Feature extractor: fenc is implemented as a 2-stack hourglass network [58] following PIFu [69]. Given the instance mask, we take out the object of interest in the picture and resize it to 256 256 before providing it as input to fenc. The output feature map has dimensions of 128 128 256 of spatial resolution 128 128 and feature dimension K = 256. Velocity-field network: v uses a 4-layer MLP with skip-connections following [59], where the internal dimension is fixed to 128. It takes query points as input and adds the visual features to the activations after the 1st block. For ODE solvers, we use the Dormand Prince method (dopri5) [12]. Training: In each training iteration we sample 2048 query points for shape reconstruction and 512 vertices for learning of temporal coherence. We train REDO end-to-end using the Adam optimizer [39] for 60 epochs with a batch size of 8. The learning rate is initialized to 0.0001 and decayed by 10 at the 40th and 55th epochs. 4.2 Experimental setup Dataset. We briefly introduce the three datasets used in our experiments below. For more details (e.g., preparation, statistics, examples), please check Appendix B. Dataset SAIL-VOS 3D Deforming Things4D++ 3DPW Metrics m Io U" m Cham.# m ACD# m Io U" m Cham.# m ACD# m Io U" m Cham.# m ACD# Static reconstruction ONet [53] 24.5 0.951 - 60.2 0.260 - 29.8 0.440 - PIFu HD [70] 25.6 0.724 - 43.8 0.511 - 37.4 0.363 - Dynamic reconstruction Surfel Warp [20] 1.03 2.13 - 3.75 6.53 - - - - OFlow [59] 26.0 0.732 1.69 55.2 0.412 0.812 31.5 0.461 0.907 REDO 31.9 0.647 1.47 57.4 0.349 0.765 41.6 0.337 0.846 Table 1: Quantitative results. For both shape reconstruction (m Io U and m Cham.) and dynamics modeling (m ACD), REDO demonstrates significant improvements over prior methods. m ACD is not available for static methods and Surfel Warp which don t predict temporally corresponding meshes. 1) SAIL-VOS 3D [27]: a photo-realistic synthetic dataset extracted from the game GTA-V. It consists of RGBD videos together with ground-truth (masks and cameras). Out of the original 178 object categories, we use 7 dynamic ones: human, car, truck, motorcycle, bicycle, airplane, and helicopter. During training, we randomly sample clips from 193 training videos. For evaluation, we sample 291 clips from 78 validation videos. We further hold out 2 classes (dog and gorilla) as an unseen test set. 2) Deforming Things4D++: Deforming Things4D [42] is a synthetic dataset containing 39 object categories. As the original dataset only provides texture-less meshes, we render RGBD video and corresponding ground-truth (mask and camera) using Blender. Because the original dataset doesn t provide dataset splits, we create our own. Specifically, during training, we randomly sample clips from 1227 videos. For evaluation, we create a validation set of 152 clips and a test set of 347 clips. We hold out class puma with 56 videos as a zero-shot test set. We dub this dataset Deforming Things4D++. 3) 3D Poses in the Wild (3DPW) [86]: to test the generalizability of our model, we test on this realworld video dataset. Unfortunately, no real-world multi-class 4D dataset is available. Therefore, we test REDO in a class-specific, i.e., class human, setting using 3DPW. This dataset contains calibrated videos, i.e., known camera, and 3D human pose annotation. However, it doesn t provide ground-truth mesh and depth. To extract a mesh, we fit the provided 3D human pose using the SMPL [47] template. To compute depth, we use Consistent Video Depth (CVD) [51] with ground-truth camera data to get temporally consistent estimates. The dataset contains 60 videos (24 training, 12 validation, and 24 testing). During training, we randomly sample clips from all training videos. For evaluation, we evaluate at uniformly sampled clips (10 clips per video) using the validation and test set. Baselines. We consider the following baselines: 1) Static reconstruction: we adopt state-of-the-art methods ONet [53] and PIFu HD [70], and train them for per-frame static reconstruction. For a fair comparison, we train these two networks in a class-agnostic setting using all the frames in the training videos. 2) Fusion-based dynamic reconstruction: most fusion based methods [57, 74, 75] are neither open-sourced nor reproduced. Among the available ones, we adapt the author-released Surfel Warp [20] due to its superior performance. Since this method is non-parametric and requires no training, we directly apply it on the validation/testing clips. 3) Supervised dynamic reconstruction: REDO learns to reconstruct dynamic objects in a supervised manner. OFlow [59] also falls into this category but it doesn t handle the partial observation and rigid-motion. Note, REDO uses each clip s center frame as the canonical space while OFlow uses the initial one. Therefore, for a fair comparison, we set OFlow s first frame to be the center of our input clip. Metrics. To evaluate the reconstructed geometry, we report the mean volumetric Intersection over Union (m Io U) and the mean Chamfer 1 distance (m Cham.) over different classes at one timestep, i.e., the center frame of the test clip. To evaluate the temporal motion prediction, we compute the mean Averaged (over time) Correspondence 2 Distance (m ACD) following [59]. As stated before, OFlow s starting frame is set to the center frame. For a fair comparison, we report m ACD on the latter half of each testing clip. We compute m Cham. and m ACD error in a scale-invariant way following [53, 59, 18]: we use 1/10 times the maximal edge length of the object s bounding box as unit one. Even though our network is class-agnostic, we report the mean values over different object categories. Namely, all mean operations are conducted over categories. 4.3 Quantitative results We present results on all three datasets in Tab. 1. For a fair comparison, we test on the center frame (canonical frame of REDO) of the validation/testing clips. We observe that: 1) REDO improves upon the static methods for shape reconstruction on SAIL-VOS 3D and 3DPW (+6.3/4.2 m Io U and -0.077/-0.026 m Cham. over best static method). This is because the static methods cannot capture the visual information from other frames in the video clip and thus fail to handle partial visibility. However, REDO performs slightly worse than ONet on Deforming Things4D++ due to the unrealistic simplified visual input: 1) the pictures have only one foreground object with neither occlusion nor background, 2) the rendered color is determined by the vertices order and hence provides a visual short-cut for 2D to 3D mapping. Without modeling dynamics, the static baselines are hence easier to optimize. In contrast, SAIL-VOS 3D renders photo-realistic game scenes with diverse dynamic objects, which are much closer to a real-world setting. 2) REDO outperforms fusion-based method Surfel Warp greatly on all benchmarks, as Surfel Warp only recovers the observable surface rather than the complete shape. We didn t run Surfel Warp on 3DPW as it relies on precise depth as input and crashes frequently using the estimated depth values. 3) REDO improves upon OFlow (+5.9/2.2/10.1 m Io U and -0.085/0.063/0.124 m Cham.) for shape reconstruction due to the pixel-aligned 4D implicit representation, whereas OFlow encodes the whole image as a single feature vector and looses spatial information. 4) Regarding dynamics modeling, REDO improves upon OFlow (-0.22/0.047/0.061 m ACD) thanks to the pixel-aligned implicit flow-field. Note that OFlow normalizes the 3D models at each time-step into the first frame s model space and hence fails to capture rigid motion like translation. In contrast, our canonical space is constructed for the entire clip in which REDO predicts a complete trajectory. As stated in Sec. 3.3, the reconstructed mesh on the center frame is propagated to other frames in the video clip for a dense reconstruction. We thus report the per-frame results in Appendix E. In addition, all above results are mean values averaged over object categories to avoid being biased towards the most frequent class. Class-wise results on SAIL-VOS 3D are reported in Appendix C. m Io U" m Cham.# m ACD# ONet 23.1 0.764 - PIFu 21.2 0.911 - Surfel Warp 2.06 1.23 - OFlow 26.7 0.931 1.18 REDO 38.5 0.479 1.07 Table 2: Zero-shot reconstruction. Zero-shot reconstruction. To test the generalizability of REDO, we further test on unseen categories with no finetuning in Tab. 2. The result is averaged over three unseen classes: dog and gorilla from SAIL-VOS 3D, and puma from Deforming Things4D++. REDO still greatly outperforms baselines and doesn t fail catastrophically. The perclass results are provided in Appendix Tab. S5. 4.4 Analysis m Io U" m Cham.# m ACD# avg. pooling 28.3 0.712 1.60 w/o alignment 24.1 0.937 1.85 w/o Ltemp 29.4 0.685 3.12 REDO 31.9 0.647 1.47 Table 3: Ablation studies. In Tab. 3, we provide an ablation study of different components in REDO using SAIL-VOS 3D data. 1) We first replace the temporal aggregator fagg with an average pooling layer where features of different frames are averaged and fed into the shape reconstruction and velocity field network. The results are shown in the 1st row of Tab. 3 (avg. pooling). The performance drops by -3.6 Io U, +0.065 m Cham., and +0.13 m ACD. 2) We then study pixel-aligned feature representations xp, zp. We replace these two features with the feature map of the entire input frame following OFlow [59] but still keep the transformer to aggregate these feature maps. Results of this ablation are reported in Tab. 3 (w/o alignment). Compared to REDO, this setting greatly hurts the results (-7.8 m Io U, +0.290 m Cham., -0.38 m ACD) as the network can no longer handle partial observations and the 3D predictions don t well align with the visual input. 3) In many real-world tasks, ground-truth meshes of different time-steps are not corresponded. Conceptually, REDO could adapt to this setting. This is because all components are differentiable and the flow-field network could be used as a latent module for shape reconstruction. To mimic this setting, we train REDO using only the shape reconstruction loss Lshape. As shown in Tab. 3 (w/o Ltemp), the model still recovers objects at the canonical frame. However, m ACD increases significantly (+1.65). 4.5 Qualitative results Fig. 4 shows a few representative examples of REDO predictions on SAIL-VOS 3D and Deforming Things4D++. Please check Appendix F for more results on real-world data and additional analysis. From Fig. 4 we observe that: 1) REDO is able to recover accurate geometry and dynamics of different objects from input video frames. It completes the occluded parts and hallucinates invisible GT Input PIFu HD Surfel Warp OFlow Ours Figure 4: Qualitative results. We illustrate the input frames with the object of interest highlighted, the reconstructed meshes obtained from different methods, and the ground-truth mesh. Figure 5: Flow-field visualization. REDO accurately recovers the non-rigid motion (e.g., moving forward) but is less precise for small-scale articulation (e.g., hand). components (e.g., legs and back of humans; rear tire of the motorcycle) by aggregating temporal information and due to large-scale training. 2) REDO improves upon baselines methods. E.g., PIFu HD struggles to handle occlusion and non-human objects, Surfel Warp only predicts the visible surface, and OFlow results are over-smooth as it ignores the spatial information. 3) REDO predictions are still far from perfect compared to ground-truth meshes. Many fine-grained details are missing (e.g., clothing of the human, car s front-light and tires, etc.). We also visualize the predicted and ground-truth motion vectors in Fig. 5. REDO successfully models the rigid motion, i.e., moving forward, over the entire human body, e.g., head, leg, chest, etc. It s less accurate in capturing the very fine-grained dynamics, e.g., the hand motion where the ground-truth indicates the hand will open while the prediction doesn t. 5 Conclusion We present REDO, a novel class-agnostic method to reconstruct dynamic objects from videos. REDO is implemented as a canonical 4D implicit function which captures the precise shape and dynamics and deals with partial visibility. We validate the effectiveness of REDO on synthetic and real-world datasets. We think REDO could be generalized to a wide variety of 4D reconstruction tasks. Acknowledgements & funding transparency statement. This work was supported in part by NSF under Grant #1718221, 2008387, 2045586, 2106825, MRI #1725729, NIFA award 2020-6702132799 and Cisco Systems Inc. (Gift Award CG 1377144 - thanks for access to Arcetri). ZR is supported by a Yee Memorial Fund Fellowship. [1] P. Achlioptas, O. Diamanti, I. Mitliagkas, and L. Guibas. Learning representations and generative models for 3d point clouds. In ICML, 2018. [2] A. Baak, M. Müller, G. Bharaj, H.-P. Seidel, and C. Theobalt. A data-driven approach for real-time full body pose reconstruction from a depth camera. Consumer Depth Cameras for Computer Vision, 2013. [3] F. Bogo, J. Romero, G. Pons-Moll, and M. J. Black. Dynamic FAUST: Registering human bodies in motion. In CVPR, 2017. [4] A. Bozic, P. Palafox, M. Zollöfer, A. Dai, J. Thies, and M. Nießner. Neural non-rigid tracking. In Neur IPS, [5] A. Božiˇc, M. Zollhöfer, C. Theobalt, and M. Nießner. Deepdeform: Learning non-rigid rgb-d reconstruction with semi-supervised data. In CVPR, 2020. [6] R. T. Chen, Y. Rubanova, J. Bettencourt, and D. Duvenaud. Neural ordinary differential equations. In Neur IPS, 2018. [7] Z. Chen and H. Zhang. Learning implicit fields for generative shape modeling. In CVPR, 2019. [8] C. B. Choy, D. Xu, J. Gwak, K. Chen, and S. Savarese. 3d-r2n2: A unified approach for single and multi-view 3d object reconstruction. In ECCV, 2016. [9] A. Collet, M. Chuang, P. Sweeney, D. Gillett, D. Evseev, D. Calabrese, H. Hoppe, A. Kirk, and S. Sullivan. High-quality streamable free-viewpoint video. TOG, 2015. [10] B. Curless and M. Levoy. A volumetric method for building complex models from range images. In SIGGRAPH, 1996. [11] E. de Aguiar, C. Stoll, C. Theobalt, N. Ahmed, H.-P. Seidel, and S. Thrun. Performance capture from sparse multi-view video. TOG, 2008. [12] J. R. Dormand and P. J. Prince. A family of embedded runge-kutta formulae. J. Comput. Appl. Math, 1980. [13] A. Dosovitskiy, P. Fischer, E. Ilg, P. Hausser, C. Hazirbas, V. Golkov, P. Van Der Smagt, D. Cremers, and T. Brox. Flownet: Learning optical flow with convolutional networks. In ICCV, 2015. [14] M. Dou, J. Taylor, H. Fuchs, A. Fitzgibbon, and S. Izadi. 3d scanning deformable objects with a single rgbd sensor. In CVPR, 2015. [15] M. Dou, S. Khamis, Y. Degtyarev, P. L. Davidson, S. Fanello, A. Kowdle, S. Orts, C. Rhemann, D. Kim, J. Taylor, P. Kohli, V. Tankovich, and S. Izadi. Fusion4d: real-time performance capture of challenging scenes. TOG, 2016. [16] M. Dou, P. Davidson, S. R. Fanello, S. Khamis, A. Kowdle, C. Rhemann, V. Tankovich, and S. Izadi. Motion2fusion: Real-time volumetric performance capture. TOG, 2017. [17] Facebook-AI. D2go brings detectron2 to mobile. https://ai.facebook.com/blog/ d2go-brings-detectron2-to-mobile/, 2021. [18] H. Fan, H. Su, and L. J. Guibas. A point set generation network for 3d object reconstruction from a single image. In CVPR, 2017. [19] C. Franklin. Apple unveils new i Pad Pro with breakthrough Li DAR Scanner and brings trackpad support to i Pad OS. https://www.apple.com/, 2020. [20] W. Gao and R. Tedrake. Surfelwarp: Efficient non-volumetric single view dynamic reconstruction. In RSS, [21] D. Gavrila and L. Davis. Tracking of humans in action: A 3-d model-based approach. ARPA Image Understanding Workshop, 1996. [22] G. Gkioxari, J. Malik, and J. Johnson. Mesh r-cnn. In ICCV, 2019. [23] T. Groueix, M. Fisher, V. G. Kim, B. Russell, and M. Aubry. Atlas Net: A Papier-Mâché Approach to Learning 3D Surface Generation. In CVPR, 2018. [24] K. Guo, F. Xu, T. Yu, X. Liu, Q. Dai, and Y. Liu. Real-time geometry, albedo, and motion reconstruction using a single rgb-d camera. TOG, 2017. [25] M. Habermann, W. Xu, M. Zollhoefer, G. Pons-Moll, and C. Theobalt. Livecap: Real-time human performance capture from monocular video. TOG, 2019. [26] K. He, G. Gkioxari, P. Dollár, and R. Girshick. Mask r-cnn. In CVPR, 2017. [27] Y.-T. Hu, J. Wang, R. A. Yeh, and A. G. Schwing. SAIL-VOS 3D: A Synthetic Dataset and Baselines for Object Detection and 3D Mesh Reconstruction from Video Data. In CVPR, 2021. [28] M. Innmann, M. Zollhöfer, M. Nießner, C. Theobald, and M. Stamminger. Volumedeform: Real-time volumetric non-rigid reconstruction. In ECCV, 2016. [29] C. Ionescu, D. Papava, V. Olaru, and C. Sminchisescu. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. TPAMI, 2013. [30] S. Izadi, D. Kim, O. Hilliges, D. Molyneaux, R. A. Newcombe, P. Kohli, J. Shotton, S. Hodges, D. Freeman, A. J. Davison, and A. W. Fitzgibbon. Kinectfusion: real-time 3d reconstruction and interaction using a moving depth camera. In UIST, 2011. [31] M. Jaimez, M. Souiai, J. González, and D. Cremers. A primal-dual framework for real-time dense rgb-d scene flow. In ICRA, 2015. [32] H. Joo, H. Liu, L. Tan, L. Gui, B. C. Nabbe, I. Matthews, T. Kanade, S. Nobuhara, and Y. Sheikh. Panoptic studio: A massively multiview system for social motion capture. In ICCV, 2015. [33] H. Joo, T. Simon, and Y. Sheikh. Total Capture: A 3D Deformation Model for Tracking Faces, Hands, and Bodies. In CVPR, 2018. [34] A. Kanazawa, M. J. Black, D. W. Jacobs, and J. Malik. End-to-end Recovery of Human Shape and Pose. In CVPR, 2018. [35] A. Kanazawa, S. Tulsiani, A. A. Efros, and J. Malik. Learning Category-Specific Mesh Reconstruction from Image Collections. In ECCV, 2018. [36] A. Kanazawa, J. Y. Zhang, P. Felsen, and J. Malik. Learning 3d human dynamics from video. In CVPR, [37] H. Kato, Y. Ushiku, and T. Harada. Neural 3d mesh renderer. In CVPR, 2018. [38] M. Kazhdan, M. Bolitho, and H. Hoppe. Poisson surface reconstruction. In Eurographics, 2006. [39] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICLR, 2015. [40] H. Li, B. Adams, L. J. Guibas, and M. Pauly. Robust single-view geometry and motion reconstruction. [41] X. Li, S. Liu, K. Kim, S. De Mello, V. Jampani, M.-H. Yang, and J. Kautz. Self-supervised single-view 3d reconstruction via semantic consistency. In ECCV, 2020. [42] Y. Li, H. Takehara, Takafumi, Taketomi, B. Zheng, and M. Nießner. 4dcomplete: Non-rigid motion estimation beyond the observable surface. In ICCV, 2021. [43] Z. Li, S. Niklaus, N. Snavely, and O. Wang. Neural Scene Flow Fields for Space-Time View Synthesis of Dynamic Scenes. In CVPR, 2021. [44] H. Liu, R. A. R. Soto, F. Xiao, and Y. J. Lee. Yolactedge: Real-time instance segmentation on the edge. In ICRA, 2021. [45] X. Liu, C. R. Qi, and L. J. Guibas. Flownet3d: Learning scene flow in 3d point clouds. In CVPR, 2019. [46] M. Loper, N. Mahmood, and M. J. Black. Mosh: Motion and shape capture from sparse markers. TOG, [47] M. Loper, N. Mahmood, J. Romero, G. Pons-Moll, and M. J. Black. SMPL: A skinned multi-person linear model. TOG, 2015. [48] M. M. Loper and M. J. Black. Opendr: An approximate differentiable renderer. In ECCV, 2014. [49] W. E. Lorensen and H. E. Cline. Marching cubes: A high resolution 3d surface construction algorithm. [50] L. Luo, H. Li, and S. Rusinkiewicz. Structure-aware hair capture. TOG, 2013. [51] X. Luo, J. Huang, R. Szeliski, K. Matzen, and J. Kopf. Consistent video depth estimation. TOG, 2020. [52] D. Maturana and S. Scherer. Voxnet: A 3d convolutional neural network for real-time object recognition. In IROS, 2015. [53] L. Mescheder, M. Oechsle, M. Niemeyer, S. Nowozin, and A. Geiger. Occupancy networks: Learning 3d reconstruction in function space. In CVPR, 2019. [54] B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020. [55] C. Nash, Y. Ganin, S. A. Eslami, and P. Battaglia. Polygen: An autoregressive generative model of 3d meshes. In ICML, 2020. [56] R. A. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A. J. Davison, P. Kohli, J. Shotton, S. Hodges, and A. W. Fitzgibbon. Kinectfusion: Real-time dense surface mapping and tracking. In ISMAR, 2011. [57] R. A. Newcombe, D. Fox, and S. M. Seitz. Dynamicfusion: Reconstruction and tracking of non-rigid scenes in real-time. In CVPR, 2015. [58] A. Newell, K. Yang, and J. Deng. Stacked hourglass networks for human pose estimation. In ECCV, 2016. [59] M. Niemeyer, L. Mescheder, M. Oechsle, and A. Geiger. Occupancy flow: 4d reconstruction by learning particle dynamics. In ICCV, 2019. [60] J. J. Park, P. R. Florence, J. Straub, R. A. Newcombe, and S. Lovegrove. Deep SDF: Learning Continuous Signed Distance Functions for Shape Representation. In CVPR, 2019. [61] K. Park, U. Sinha, J. T. Barron, S. Bouaziz, D. B. Goldman, S. M. Seitz, and R. Martin-Brualla. Nerfies: Deformable neural radiance fields. In ICCV, 2021. [62] S. I. Park and J. K. Hodgins. Capturing and animating skin deformation in human motion. TOG, 2006. [63] A. Pumarola, E. Corona, G. Pons-Moll, and F. Moreno-Noguer. D-Ne RF: Neural radiance fields for dynamic scenes. In CVPR, 2020. [64] C. R. Qi, H. Su, M. Nießner, A. Dai, M. Yan, and L. J. Guibas. Volumetric and multi-view cnns for object classification on 3d data. In CVPR, 2016. [65] C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In CVPR, 2017. [66] C. R. Qi, L. Yi, H. Su, and L. J. Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Neur IPS, 2017. [67] Z. Ren, I. Misra, A. G. Schwing, and R. Girdhar. 3d spatial recognition without spatially labeled 3d. In CVPR, 2021. [68] G. Riegler, A. Osman Ulusoy, and A. Geiger. Octnet: Learning deep 3d representations at high resolutions. In CVPR, 2017. [69] S. Saito, Z. Huang, R. Natsume, S. Morishima, A. Kanazawa, and H. Li. PIFu: Pixel-Aligned Implicit Function for High-Resolution Clothed Human Digitization. In ICCV, 2019. [70] S. Saito, T. Simon, J. Saragih, and H. Joo. PIFu HD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization. In CVPR, 2020. [71] J. L. Schönberger and J.-M. Frahm. Structure-from-motion revisited. In CVPR, 2016. [72] T. Simon, H. Joo, and Y. Sheikh. Hand keypoint detection in single images using multiview bootstrapping. In CVPR, 2017. [73] V. Sitzmann, M. Zollhöfer, and G. Wetzstein. Scene representation networks: Continuous 3d-structure- aware neural scene representations. In Neur IPS, 2019. [74] M. Slavcheva, M. Baust, D. Cremers, and S. Ilic. Killing Fusion: Non-rigid 3D Reconstruction without Correspondences. In CVPR, 2017. [75] M. Slavcheva, M. Baust, and S. Ilic. Sobolevfusion: 3d reconstruction of scenes undergoing free non-rigid motion. In CVPR, 2018. [76] S. Song and J. Xiao. Deep sliding shapes for amodal 3d object detection in rgb-d images. In CVPR, 2016. [77] J. Starck and A. Hilton. Surface capture for performance-based animation. CG&A, 2007. [78] S. Stein. Lidar on the i Phone 12 Pro. https://www.cnet.com/, 2020. [79] Z. Su, W. Wan, T. Yu, L. Liu, L. Fang, W. Wang, and Y. Liu. Mulaycap: Multi-layer human performance capture using a monocular video camera. TVCG, 2020. [80] D. Sun, X. Yang, M.-Y. Liu, and J. Kautz. Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In CVPR, 2018. [81] T. Takikawa, J. Litalien, K. Yin, K. Kreis, C. Loop, D. Nowrouzezahrai, A. Jacobson, M. Mc Guire, and S. Fidler. Neural geometric level of detail: Real-time rendering with implicit 3D shapes. In CVPR, 2021. [82] M. Tatarchenko, A. Dosovitskiy, and T. Brox. Octree generating networks: Efficient convolutional architectures for high-resolution 3d outputs. In ICCV, 2017. [83] Z. Teed and J. Deng. Raft: Recurrent all-pairs field transforms for optical flow. In ECCV, 2020. [84] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin. Attention is all you need. In Neur IPS, 2017. [85] D. Vlasic, I. Baran, W. Matusik, and J. Popovi c. Articulated mesh animation from multi-view silhouettes. [86] T. von Marcard, R. Henschel, M. Black, B. Rosenhahn, and G. Pons-Moll. Recovering accurate 3d human pose in the wild using imus and a moving camera. In ECCV, 2018. [87] T. von Marcard, R. Henschel, M. J. Black, B. Rosenhahn, and G. Pons-Moll. Recovering accurate 3d human pose in the wild using imus and a moving camera. In ECCV, 2018. [88] N. Wang, Y. Zhang, Z. Li, Y. Fu, W. Liu, and Y.-G. Jiang. Pixel2mesh: Generating 3d mesh models from single rgb images. In ECCV, 2018. [89] T. Weise, S. Bouaziz, H. Li, and M. Pauly. Realtime performance-based facial animation. TOG, 2011. [90] J. Wu, C. Zhang, T. Xue, W. T. Freeman, and J. B. Tenenbaum. Learning a probabilistic latent space of object shapes via 3d generative-adversarial modeling. In Neur IPS, 2016. [91] G. Yang, D. Sun, V. Jampani, D. Vlasic, F. Cole, H. Chang, D. Ramanan, W. T. Freeman, and C. Liu. Lasr: Learning articulated shape reconstruction from a monocular video. In CVPR, 2021. [92] T. Yu, K. Guo, F. Xu, Y. Dong, Z. Su, J. Zhao, J. Li, Q. Dai, and Y. Liu. Bodyfusion: Real-time capture of human motion and surface geometry using a single depth camera. In ICCV, 2017. [93] T. Yu, Z. Zheng, K. Guo, J. Zhao, Q. Dai, H. Li, G. Pons-Moll, and Y. Liu. Doublefusion: Real-time capture of human performances with inner body shapes from a single depth sensor. In CVPR, 2018. [94] J. Y. Zhang, P. Felsen, A. Kanazawa, and J. Malik. Predicting 3d human dynamics from video. In ICCV, [95] M. Zollhöfer, M. Nießner, S. Izadi, C. Rhemann, C. Zach, M. Fisher, C. Wu, A. W. Fitzgibbon, C. T. Loop, C. Theobalt, and M. Stamminger. Real-time non-rigid reconstruction using an rgb-d camera. TOG, 2014. [96] S. Zuffi, A. Kanazawa, and M. J. Black. Lions and tigers and bears: Capturing non-rigid 3d articulated shape from images. In CVPR, 2018. [97] S. Zuffi, A. Kanazawa, T. Berger-Wolf, and M. J. Black. Three-D Safari: Learning to Estimate Zebra Pose, Shape, and Texture from Images" In the Wild". In ICCV, 2019.