# neural_nonrigid_tracking__4e6ff229.pdf Neural Non-Rigid Tracking Aljaž Božiˇc1 aljaz.bozic@tum.de Pablo Palafox1 pablo.palafox@tum.de Michael Zollhöfer2 Angela Dai1 Justus Thies1 Matthias Nießner1 1Technical University of Munich 2Stanford University We introduce a novel, end-to-end learnable, differentiable non-rigid tracker that enables state-of-the-art non-rigid reconstruction by a learned robust optimization. Given two input RGB-D frames of a non-rigidly moving object, we employ a convolutional neural network to predict dense correspondences and their confidences. These correspondences are used as constraints in an as-rigid-as-possible (ARAP) optimization problem. By enabling gradient back-propagation through the weighted non-linear least squares solver, we are able to learn correspondences and confidences in an end-to-end manner such that they are optimal for the task of nonrigid tracking. Under this formulation, correspondence confidences can be learned via self-supervision, informing a learned robust optimization, where outliers and wrong correspondences are automatically down-weighted to enable effective tracking. Compared to state-of-the-art approaches, our algorithm shows improved reconstruction performance, while simultaneously achieving 85 faster correspondence prediction than comparable deep-learning based methods. We make our code available at https://github.com/Deformable Friends/Neural Tracking. 1 Introduction The capture and reconstruction of real-world environments is a core problem in computer vision, enabling numerous VR/AR applications. While there has been significant progress in reconstructing static scenes, tracking and reconstruction of dynamic objects remains a challenge. Non-rigid reconstruction focuses on dynamic objects, without assuming any explicit shape priors, such as human or face parametric models. Commodity RGB-D sensors, such as Microsoft s Kinect or Intel s Realsense, provide a cost-effective way to acquire both color and depth video of dynamic motion. Using a large number of RGB-D sensors can lead to an accurate non-rigid reconstruction, as shown by Dou et al. [8]. Our work focuses on non-rigid reconstruction from a single RGB-D camera, thus eliminating the need for specialized multi-camera setups. The seminal Dynamic Fusion by Newcombe et al. [23] introduced a non-rigid tracking and mapping pipeline that uses depth input for real-time non-rigid reconstruction from a single RGB-D camera. Various approaches have expanded upon this framework by incorporating sparse color correspondences [13] or dense photometric optimization [10]. Deep Deform [4] presented a learned correspondence prediction, enabling significantly more robust tracking of fast motion and re-localization. Unfortunately, the computational cost of the correspondence prediction network ( 2 seconds per frame for a relatively small number of non-rigid correspondences) inhibits real-time performance. Denotes equal contribution. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Depth Integration Correspondence Correspondence RGB-D Input Stream Source Target Neural Non-Rigid Tracker Final Reconstruction Differentiable Figure 1: Neural Non-Rigid Tracking: based on RGB-D input data of a source and a target frame, our learned non-rigid tracker estimates the non-rigid deformations to align the source to the target frame. We propose an end-to-end approach, enabling correspondences and their importance weights to be informed by the non-rigid solver. Similar to robust optimization, this provides robust tracking, and the resulting deformation field can then be used to integrate the depth observations in a canonical volumetric 3D grid that implicitly represents the surface of the object (final reconstruction). Simultaneously, work on learned optical flow has shown dense correspondence prediction at real-time rates [30]. However, directly replacing the non-rigid correspondence predictions from Božiˇc et al. [4] with these optical flow predictions does not produce accurate enough correspondences for comparable non-rigid reconstruction performance. In our work, we propose a neural non-rigid tracker, i.e., an end-to-end differentiable non-rigid tracking pipeline which combines the advantages of classical deformation-graph-based reconstruction pipelines [23, 13] with novel learned components. Our end-to-end approach enables learning outlier rejection in a self-supervised manner, which guides a robust optimization to mitigate the effect of inaccurate correspondences or major occlusions present in single RGB-D camera scenarios. Specifically, we cast the non-rigid tracking problem as an as-rigid-as-possible (ARAP) optimization problem, defined on correspondences between points in a source and a target frame. A differentiable Gauss-Newton solver allows us to obtain gradients that enable training a neural network to predict an importance weight for every correspondence in a completely self-supervised manner, similar to robust optimization. The end-to-end training significantly improves non-rigid tracking performance. Using our neural tracker in a non-rigid reconstruction framework results in 85 faster correspondence prediction and improved reconstruction performance compared to the state of the art. In summary, we propose a novel neural non-rigid tracking approach with two key contributions: an end-to-end differentiable Gauss-Newton solver, which provides gradients to better inform a correspondence prediction network used for non-rigid tracking of two frames; a self-supervised approach for learned correspondence weighting, which is informed by our differentiable solver and enables efficient, robust outlier rejection, thus, improving non-rigid reconstruction performance compared to the state of the art. 2 Related Work Non-rigid Reconstruction. Reconstruction of deformable objects using a single RGB-D camera is an important research area in computer vision. State-of-the-art methods rely on deformation graphs [29, 35] that enable robust and temporally consistent 3D motion estimation. While earlier approaches required an object template, such graph-based tracking has been extended to simultaneous tracking and reconstruction approaches [7, 23]. These works used depth fitting optimization objectives in the form of iterative closest points, or continuous depth fitting in [26, 27]. Rather than relying solely on depth information, recent works have incorporated SIFT features [13], dense photometric fitting [10], and sparse learned correspondence [4]. Correspondence Prediction for Non-rigid Tracking. In non-rigid tracking, correspondences must be established between the two frames we want to align. While methods such as Dynamic Fusion [23] rely on projective correspondences, recent methods leverage learned correspondences [4]. Deep Deform [4] relies on sparse predicted correspondences, trained on an annotated dataset of deforming objects. Since prediction is done independently for each correspondence, this results in a high compute cost, compared to dense predictions of state-of-the-art optical flow networks. Optical flow [6, 12, 30, 18] and scene flow [21, 3, 19, 33] methods achieve promising results in predicting dense correspondences between two frames, with some approaches not even requiring direct supervision [32, 16, 15]. In our proposed neural non-rigid tracking approach, we build upon PWC-Net [30] for dense correspondence prediction to inform our non-rigid deformation energy formulation. Since our approach allows for end-to-end training, our 2D correspondence prediction finds correspondences better suited for non-rigid tracking. Differentiable Optimization. Differentiable optimizers have been explored for various tasks, including image alignment [5], rigid pose estimation [11, 20], multi-frame direct bundleadjustment [31], and rigid scan-to-CAD alignment [1]. In addition to achieving higher accuracy, an end-to-end differentiable optimization approach also offers the possibility to optimize run-time, as demonstrated by learning efficient pre-conditioning methods in [9, 25, 17]. Unlike Li et al. [17], which employs an image-based tracker (with descriptors defined on nodes in a pixel-aligned graph), our approach works on general graphs and learns to robustify correspondence prediction for non-rigid tracking by learning self-supervised correspondence confidences. 3 Non-Rigid Reconstruction Notation Non-rigid alignment is a crucial part of non-rigid reconstruction pipelines. In the single RGB-D camera setup, we are given a pair of source and target RGB-D frames {(Is, Ps), (It, Pt)}, where I RH W 3 is an RGB image and P RH W 3 a 3D point image. The goal is to estimate a warp field Q : R3 7 R3 that transforms Ps into the target frame. Note that we define the 3D point image Ps as the result of back-projecting every pixel u Πs R2 into the camera coordinate system with given camera intrinsic parameters. To this end, we define the inverse of the perspective projection to back-project a pixel u given the pixel s depth du and the intrinsic camera parameters c: π 1 c : R2 R R3, (u, du) 7 π 1 c (u, du) = p. (1) To maintain robustness against noise in the depth maps, state-of-the-art approaches define an embedded deformation graph G = {V, E} over the source RGB-D frame, where V is the set of graph nodes defined by their 3D coordinates vi R3 and E the set of edges between nodes, as described in [29] and illustrated in Fig. 1. Thus, for every node in G, a global translation vector tvi R3 and a rotation matrix Rvi R3 3, must be estimated in the alignment process. We parameterize rotations with a 3-dimensional axis-angle vector ω R3. We use the exponential map exp : so(3) SO(3), bω 7 ebω = R to convert from axis-angle to matrix rotation form, where the b -operator creates a 3 3 skew-symmetric matrix from a 3-dimensional vector. The resulting graph motion is denoted by T = (ωv1, tv1, . . . , ωv N , tv N ) RN 6 for a graph with N nodes. Dense motion can then be computed by interpolating the nodes motion T by means of a warping function Q. When applied to a 3D point p R3, it produces the point s deformed position Q(p, T ) = X vi V αvi(ebωvi(p vi) + vi + tvi). (2) The weights αvi R, also known as skinning weights, measure the influence of each node on the current point p and are computed as in [34]. Please see the supplemental material for further detail. 4 Neural Non-rigid Tracking Given a pair of source and target RGB-D frames (Zs, Zt), where Z = (I |P ) RH W 6 is the concatenation of an RGB and a 3D point image as defined in Section 3, we aim to find a function Θ that estimates the motion T of a deformation graph G with N nodes (given by their 3D coordinates V) defined over the source RGB-D frame. This implicitly defines source-to-target dense 3D motion (see Figure 2). Formally, we have: Θ : RH W 6 RH W 6 RN 3 RN 6, (Zs, Zt, V) 7 Θ (Zs, Zt, V) = T . (3) Correspondence Differentiable Solver Correspondence It Prediction Sample Graph Figure 2: Overview of our neural non-rigid tracker. Given a pair of source and target images, Is and It, a dense correspondence map C between the frames is estimated via a convolutional neural network Φ. Importance weights W for these correspondences are computed through a function Ψ. Together with a graph G defined over the source RGB-D frame Ps, both C and W are input to a differentiable solver Ω. The solver outputs the graph motion T , i.e., the non-rigid alignment between source and target frames. Our approach is optimized end-to-end, with losses on the final alignment using Lgraph and Lwarp, and an intermediate loss on the correspondence map Lcorr. To estimate T , we first establish dense 2D correspondences between the source and target frame using a deep neural network Φ. These correspondences, denoted as C, are used to construct the data term in our non-rigid alignment optimization. Since the presence of outlier correspondence predictions has a strong negative impact on the performance of non-rigid tracking, we introduce a weighting function Ψ, inspired by robust optimization, to down-weight inaccurate predictions. Function Ψ outputs importance weights W and is learned in a self-supervised manner. Finally, both correspondence predictions C and importance weights W are input to a differentiable, non-rigid alignment optimization module Ω. By optimizing the non-rigid alignment energy (see Section 4.3), the differentiable optimizer Ωestimates the deformation graph parameters T that define the motion from source to target frame: T = Θ (Zs, Zt, V) = Ω(Φ( ), Ψ( ), V) = Ω(C, W, V) . (4) In the following, we define the dense correspondence predictor Φ, the importance weighting Ψ and the optimizer Ω, and describe a fully differentiable approach for optimizing Φ and Ψ such that we can estimate dense correspondences with importance weights best suited for non-rigid tracking. 4.1 Dense Correspondence Prediction The dense correspondence prediction function Φ takes as input a pair of source and target RGB images (Is, It), and for each source pixel location u Πs R2 it outputs a corresponding pixel location in the target image It, which we denote by cu Πt R2. Formally, Φ is defined as Φ : RH W 3 RH W 3 RH W 2, (Is, It) 7 Φ (Is, It) = C, (5) where C is the resulting dense correspondence map. The function Φ is represented by a deep neural network that leverages the architecture of a state-of-the-art optical flow estimator [30]. 4.2 Correspondence Importance Weights For each source pixel u Πs R2 and its correspondence cu Πt R2, we additionally predict an importance weight wu (0, 1) by means of the weighting function Ψ. The latter takes as input the source RGB-D image Zs, the corresponding sampled target frame values Z t, and intermediate features from the correspondence network Φ, and outputs weights for the correspondences between source and target. Note that Z t is the result of bilinearly sampling [14] the target image Zt at the predicted correspondence locations C. The last layer of features H of the correspondence network Φ, with dimension D = 565, are used to inform Ψ. The weighting function is thus defined as Ψ : RH W 6 RH W 6 RH W D RH W 1, (Zs, Z t, H) 7 Ψ (Zs, Z t, H) = W. (6) 4.3 Differentiable Optimizer We introduce a differentiable optimizer Ωto estimate the deformation graph parameters T , given the correspondence map C, importance weights W, and N graph nodes V: Ω: RH W 2 RH W 1 RN 3 RN 6, (C, W, V) 7 Ω(C, W, V) = T , (7) with C and W estimated by functions Φ (Eq. 5) and Ψ (Eq. 6), respectively. Using the predicted dense correspondence map C, we establish the data term for the non-rigid tracking optimization. Specifically, we use a 2D data term that operates in image space and a depth data term that leverages the depth information of the input frames. In addition to the data terms, we employ an As-Rigid-As-Possible regularizer [28] to encourage node deformations to be locally rigid, enabling robust deformation estimates even in the presence of noisy input cues. Note that the resulting optimizer module Ωis fully differentiable, but contains no learnable parameters. In summary, we formulate non-rigid tracking as the following nonlinear optimization problem: λ2DE2D(T ) + λdepth Edepth(T ) + λreg Ereg(T ) . (8) 2D reprojection term. Given the outputs of the dense correspondence predictor and weighting function, Φ (Is, It) and Ψ (Zs, Z t, H), respectively, we query for every pixel u in the source frame its correspondence cu and weight wu to build the following energy term: E2D(T ) = X u Πs w2 u πc(Q(pu, T )) cu 2 2 , (9) where πc : R3 R2, p 7 πc(p) is a perspective projection with intrinsic parameters c and pu = π 1 c (u, du) as defined in Eq. 1. Each pixel is back-projected to 3D, deformed using the current graph motion estimate as described in Eq. 2 and projected onto the target image plane. The projected deformed location is compared to the predicted correspondence cu. Depth term. The depth term leverages the depth cues of the source and target images. Specifically, it compares the z components of a warped source point, i.e., [Q(pu, T )]z, and a target point sampled at the corresponding location cu using bilinear interpolation: Edepth(T ) = X u Πs w2 u [Q(pu, T )]z [Pt(cu)]z 2. (10) Regularization term. We encourage the deformation of neighboring nodes in the deformation graph to be locally rigid. Each node vi V has at most K = 8 neighbors in the set of edges E, computed as nearest nodes using geodesic distances. The regularization term follows [28]: Ereg(T ) = X ebωvi(vj vi) + vi + tvi (vj + tvj) 2 Equation 8 is minimized using the Gauss-Newton algorithm, as described in Algorithm 1. In the following, we denote the number of correspondences by |C| and the number of graph edges by |E|. Moreover, we transform all energy terms into a residual vector r R3|C|+3|E|. For every graph node, we compute partial derivatives with respect to translation and rotation parameters, constructing a Jacobian matrix J R(3|C|+3|E|) 6N, where N is the number of nodes in the set of vertices V. Analytic formulas for partial derivatives are described in the supplemental material. Initially, the deformation parameters are initialized to T0 = 0, corresponding to zero translation and identity rotations. In each iteration n, the residual vector rn and the Jacobian matrix Jn are computed using the current estimate Tn, and the following linear system is solved (using LU decomposition) to compute an increment T : JT n Jn T = JT nrn. (12) At the end of every iteration, the motion estimate T is updated as Tn+1 = Tn + T . Most operations are matrix-matrix or matrix-vector multiplications, which are trivially differentiable. Derivatives of the linear system solve operation are computed analytically, as described in [2] and detailed in the supplement. We use max_iter = 3 Gauss-Newton iterations, which encourages the correspondence prediction and weight functions, Φ and Ψ, respectively, to make predictions such that convergence in 3 iterations is possible. In our experiments we use (λ2D, λdepth, λreg) = (0.001, 1, 1). Algorithm 1 Gauss-Newton Optimization 1: C Φ (Is, It) Estimate correspondences 2: W Ψ (Zs, Z t, H) Estimate importance weights 3: function SOLVER(C, W, V) 4: T 0 5: for n 0 to max_iter do 6: J, r Compute Jacobian And Residual(V, T , Zs, Z t, C, W) 7: T LUDecomposition(JT J T = JT r) Solve linear system 8: T T + T Apply increment 9: return T 4.4 End-to-end Optimization Given a dataset of samples Xs,t = {[Is|Ps], [It|Pt], V}, our goal is to find the parameters φ and ψ of Φφ and Ψψ, respectively, so as to estimate the motion T of a deformation graph G defined over the source RGB-D frame. This can be formulated as a differentiable optimization problem (allowing for back-propagation) with the following objective: arg min φ,ψ Xs,t λcorr Lcorr(φ) + λgraph Lgraph(φ, ψ) + λwarp Lwarp(φ, ψ) (13) Correspondence loss. We use a robust q-norm as in [30] to enforce closeness of correspondence predictions to ground-truth: Lcorr(φ) = M C( |Φφ (Is, It) C| + ϵ)q. (14) Operator | | denotes the ℓ1 norm, q < 1 (we set it to q = 0.4) and ϵ is a small constant. Ground-truth correspondences are denoted by C. Since valid ground truth for all pixels is not available, we employ a ground-truth mask M C to avoid propagating gradients through invalid pixels. Graph loss. We impose an l2-loss on node translations t (ground-truth rotations are not available): Lgraph(φ, ψ) = M V Ω Φφ (Is, It) , Ψψ (Zs, Z t, H) , V where [ ]t : RN 6 RN 3, T 7 [T ]t = t extracts the translation part from the graph motion T . Node translation ground-truth is denoted by t and M V masks out invalid nodes. Please see the supplement for further details on how M V is computed. Warp loss. We have found that it is beneficial to use the estimated graph deformation T to deform the dense source point cloud Ps and enforce the result to be close to the source point cloud when deformed with the ground-truth scene flow S: Lwarp(φ, ψ) = M S Q Ps, Ω Φφ (Is, It) , Ψψ (Zs, Z t, H) , V Here, we extend the warping operation Q (Eq. 2) to operate on the dense point cloud Ps element-wise, and define M S to mask out invalid points. Note that We found it to be a more general notation to disentangle them (e.g., for scenarios where graph nodes are not sampled on the RGB-D frame). 4.5 Neural Non-rigid Tracking for 3D Reconstruction We introduce our differentiable tracking module into the non-rigid reconstruction framework of Newcombe et al. [23]. In addition to the dense depth ICP correspondences employed in the original method, which help towards local deformation refinement, we employ a keyframe-based tracking objective. Without loss of generality, every 50th frame of the sequence is chosen as a keyframe, to Dynamic Fusion Deep Deform Ours Input Dynamic Fusion Deep Deform Ours Input Figure 3: Qualitative comparison of our method with Dynamic Fusion [23] and Deep Deform [4] on test sequences from [4]. The rows show different time steps of the sequence. which we establish dense correspondences including the respective weights. We apply a conservative filtering of the predicted correspondences based on the predicted correspondence weights using a fixed threshold δ = 0.35 and re-weight the correspondences based on bi-directional consistency, i.e., keyframe-to-frame and frame-to-keyframe. Using the correspondence predictions and correspondence weights of valid keyframes (> 50% valid correspondences), the non-rigid tracking optimization problem is solved. The resulting deformation field is used to integrate the depth frame into the canonical volume of the object. We refer to the original reconstruction paper [23] for details regarding the fusion process. 5 Experiments In the following, we evaluate our method quantitatively and qualitatively on both non-rigid tracking and non-rigid reconstruction. To this end, we use the Deep Deform dataset [4] for training, with the given 340-30-30 train-val-test split of RGB-D sequences. Both non-rigid tracking and reconstruction are evaluated on the hidden test set of the Deep Deform benchmark. 5.1 Training Scheme The non-rigid tracking module has been implemented using the Py Torch library [24] and trained using stochastic gradient descent with momentum 0.9 and learning rate 10 5. We use an Intel Xeon 6240 Processor and an Nvidia RTX 2080Ti GPU. The parameters of the dense correspondence prediction network φ are initialized with a PWC-Net model pre-trained on Flying Chairs [6] and Flying Things3D [22]. We use a 10-factor learning rate decay every 10k iterations, requiring Table 1: We evaluate non-rigid tracking on the Deep Deform dataset [4], showing the benefit of end-to-end differentiable optimizer losses and self-supervised correspondence weighting. We denote correspondence prediction as Φc, Φc+g and Φc+g+w, depending on which losses Lcorr, Lgraph, Lwarp are used, and correspondence weighting as Ψsupervised and Ψself-supervised, either using an additional supervised loss or not. Model EPE 3D (mm) Graph Error 3D (mm) Φc 44.05 67.25 Φc+g 39.12 57.34 Φc+g+w 36.96 54.24 Φc + Ψsupervised 28.95 36.77 Φc+g+w + Ψsupervised 27.42 34.68 Φc+g+w + Ψself-supervised 26.29 31.00 about 30k iterations in total for convergence, with a batch size of 4. For optimal performance, we first optimize the correspondence predictor Φφ with (λcorr, λgraph, λwarp) = (5, 5, 5), without the weighting function Ψψ. Afterwards, we optimize the weighting function parameters ψ with (λcorr, λgraph, λwarp) = (0, 1000, 1000), while keeping φ fixed. Finally, we fine-tune both φ and ψ together, with (λcorr, λgraph, λwarp) = (5, 5, 5). 5.2 Non-rigid Tracking Evaluation For any frame pair Xs,t in the Deep Deform data [4], we define a deformation graph G by uniformly sampling graph nodes V over the source object in the RGB-D frame, given a segmentation mask of the former. Graph node connectivity E is computed using geodesic distances on a triangular mesh defined over the source depth map. As a pre-processing step, we filter out any frame pairs where more than 30% of the source object is occluded in the target frame. In Table 1 non-rigid tracking performance is evaluated by the mean translation error over node translations t (Graph Error 3D), where the latter are compared to ground-truth with an l2 metric. In addition, we evaluate the dense end-point-error (EPE 3D) between the source point cloud deformed with the estimated graph motion, Q(Ps, T ), and the source point cloud deformed with the ground-truth scene flow, Ps + S. To support reproducibility, we report the mean error metrics of multiple experiments, running every setting 3 times. We visualize the standard deviation with an error plot in the supplement. We show that using graph and warp losses, Lgraph and Lwarp, and differentiating through the nonrigid optimizer considerably improves both EPE 3D and Graph Error 3D compared to only using the correspondence loss Lcorr. Adding self-supervised correspondence weighting further decreases the errors by a large margin. Supervised outlier rejection with binary cross-entropy loss does bring an improvement compared to models that do not optimize for the weighting function Ψψ (please see supplemental material for details on this supervised training of Ψψ). However, optimizing Ψψ in a self-supervised manner clearly outperforms the former supervised setup. This is due to the fact that, in the self-supervised scenario, gradients that flow from Lgraph and Lwarp through the differentiable solver Ωcan better inform the optimization of Ψψ by minimizing the end-to-end alignment losses. 5.3 Non-rigid Reconstruction Evaluation We evaluate the performance of our non-rigid reconstruction approach on the Deep Deform benchmark [4] (see Table 2). The evaluation metrics measure deformation error, a 3D end-point-error between tracked and annotated correspondences, and geometry error, which compares reconstructed shapes with annotated foreground object masks. Our approach performs about 8.9% better than the stateof-the-art non-rigid reconstruction approach of Božiˇc et al. [4] on the deformation metric. While our approach consistently shows better performance on both metrics, we also significantly lower the per-frame runtime to 27 ms per keyframe, in contrast to [4], which requires 2299 ms. Thus, our approach can also be used with multiple keyframes at interactive frames rates, e.g., 90 ms for 5 keyframes and 199 ms for 10 keyframes. Table 2: Our method achieves state-of-the-art non-rigid reconstruction results on the Deep Deform benchmark [4]. Both our end-to-end differentiable optimizer and the self-supervised correspondence weighting are necessary for optimal performance. Not only does our approach achieve lower deformation and geometry error compared to state of the art, our correspondence prediction is about 85 faster. Method Deformation error (mm) Geometry error (mm) Dynamic Fusion [23] 61.79 10.78 Volume Deform [13] 208.41 74.85 Deep Deform [4] 31.52 4.16 Ours (Φc) 54.85 5.92 Ours (Φc+g+w) 53.27 5.84 Ours (Φc + Ψsupervised) 40.21 5.39 Ours (Φc+g+w + Ψself-supervised) 28.72 4.03 To show the influence of the different learned components of our method, we perform an ablation study by disabling either of our two main components: the end-to-end differentiable optimizer or the self-supervised correspondence weighting. As can be seen, our end-to-end trained method with self-supervised correspondence weighting demonstrates the best performance. Qualitatively, we show this in Figure 3. In contrast to Dynamic Fusion [23] and Deep Deform [4], our method is notably more robust in fast motion scenarios. Additional qualitative results and comparisons to the methods of Guo et al. [10] and Slavcheva et al. [26] are shown in the supplemental material. 6 Conclusion We propose Neural Non-Rigid Tracking, a differentiable non-rigid tracking approach that allows learning the correspondence prediction and weighting of traditional tracking pipelines in an end-toend manner. The differentiable formulation of the entire tracking pipeline enables back-propagation to the learnable components, guided by a loss on the tracking performance. This not only achieves notably improved tracking error in comparison to state-of-the-art tracking approaches, but also leads to better reconstructions, when integrated into a reconstruction framework like Dynamic Fusion [23]. We hope that this work inspires further research in the direction of neural non-rigid tracking and believe that it is a stepping stone towards fully differentiable non-rigid reconstruction. Broader Impact Our paper presents learned non-rigid tracking. It is establishing the basis for the important research field of non-rigid tracking and reconstruction, which is needed for a variety of applications where man-machine and machine-environment interaction is required. These applications range from the field of augmented and virtual reality to autonomous driving and robot control. In the former, a precise understanding of dynamic and deformable objects is of major importance in order to provide an immersive experience to the user. Applications such as holographic calls would greatly benefit from research like ours. This, in turn, could provide society with the next generation of 3D communication tools. On the other hand, as a low-level building block, our work has no direct negative outcome, other than what could arise from the aforementioned applications. Acknowledgments and Disclosure of Funding This work was supported by the ZD.B (Zentrum Digitalisierung.Bayern), the Max Planck Center for Visual Computing and Communications (MPC-VCC), a TUM-IAS Rudolf Mößbauer Fellowship, the ERC Starting Grant Scan2CAD (804724), and the German Research Foundation (DFG) Grant Making Machine Learning on Static and Dynamic 3D Data Practical. [1] A. Avetisyan, A. Dai, and M. Nießner. End-to-end cad model retrieval and 9dof alignment in 3d scans. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2551 2560, 2019. [2] J. T. Barron and B. Poole. The fast bilateral solver. In European Conference on Computer Vision (ECCV), pages 617 632. Springer, 2016. [3] A. Behl, D. Paschalidou, S. Donné, and A. Geiger. Pointflownet: Learning representations for rigid motion estimation from point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 7962 7971, 2019. [4] A. Božiˇc, M. Zollhöfer, C. Theobalt, and M. Nießner. Deepdeform: Learning non-rigid rgb-d reconstruction with semi-supervised data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. [5] C.-H. Chang, C.-N. Chou, and E. Y. Chang. Clkn: Cascaded lucas-kanade networks for image alignment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2213 2221, 2017. [6] A. Dosovitskiy, P. Fischer, E. Ilg, P. Hausser, C. Hazirbas, V. Golkov, P. Van Der Smagt, D. Cremers, and T. Brox. Flownet: Learning optical flow with convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2758 2766, 2015. [7] M. Dou, J. Taylor, H. Fuchs, A. Fitzgibbon, and S. Izadi. 3d scanning deformable objects with a single rgbd sensor. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 493 501, 2015. [8] M. Dou, S. Khamis, Y. Degtyarev, P. Davidson, S. R. Fanello, A. Kowdle, S. O. Escolano, C. Rhemann, D. Kim, J. Taylor, et al. Fusion4d: Real-time performance capture of challenging scenes. ACM Transactions on Graphics (TOG), 35(4):114, 2016. [9] M. Götz and H. Anzt. Machine learning-aided numerical linear algebra: Convolutional neural networks for the efficient preconditioner generation. In 2018 IEEE/ACM 9th Workshop on Latest Advances in Scalable Algorithms for Large-Scale Systems (scal A), pages 49 56, 2018. [10] K. Guo, F. Xu, T. Yu, X. Liu, Q. Dai, and Y. Liu. Real-time geometry, albedo, and motion reconstruction using a single rgb-d camera. ACM Transactions on Graphics (TOG), 36(3):32, 2017. [11] L. Han, M. Ji, L. Fang, and M. Nießner. Regnet: Learning the optimization of direct image-to-image pose registration. ar Xiv preprint ar Xiv:1812.10212, 2018. [12] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox. Flownet 2.0: Evolution of optical flow estimation with deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2462 2470, 2017. [13] M. Innmann, M. Zollhöfer, M. Nießner, C. Theobalt, and M. Stamminger. Volumedeform: Real-time volumetric non-rigid reconstruction. In European Conference on Computer Vision, pages 362 379. Springer, 2016. [14] M. Jaderberg, K. Simonyan, A. Zisserman, and k. kavukcuoglu. Spatial transformer networks. In Advances in Neural Information Processing Systems (Neur IPS) 28, pages 2017 2025, 2015. [15] Z. Lai, E. Lu, and W. Xie. Mast: A memory-augmented self-supervised tracker. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6479 6488, 2020. [16] X. Li, S. Liu, S. De Mello, X. Wang, J. Kautz, and M.-H. Yang. Joint-task self-supervised learning for temporal correspondence. In Advances in Neural Information Processing Systems, pages 318 328, 2019. [17] Y. Li, A. Božiˇc, T. Zhang, Y. Ji, T. Harada, and M. Nießner. Learning to optimize non-rigid tracking. ar Xiv preprint ar Xiv:2003.12230, 2020. [18] P. Liu, M. Lyu, I. King, and J. Xu. Selflow: Self-supervised learning of optical flow. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4571 4580, 2019. [19] X. Liu, C. R. Qi, and L. J. Guibas. Flownet3d: Learning scene flow in 3d point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 529 537, 2019. [20] Z. Lv, F. Dellaert, J. M. Rehg, and A. Geiger. Taking a deeper look at the inverse compositional algorithm. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4581 4590, 2019. [21] W.-C. Ma, S. Wang, R. Hu, Y. Xiong, and R. Urtasun. Deep rigid instance scene flow. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3614 3622, 2019. [22] N. Mayer, E. Ilg, P. Hausser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4040 4048, 2016. [23] R. A. Newcombe, D. Fox, and S. M. Seitz. Dynamicfusion: Reconstruction and tracking of non-rigid scenes in real-time. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 343 352, 2015. [24] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. De Vito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems (Neur IPS) 32, pages 8024 8035, 2019. [25] J. Sappl, L. Seiler, M. Harders, and W. Rauch. Deep learning of preconditioners for conjugate gradient solvers in urban water related problems. ar Xiv preprint ar Xiv:1906.06925, 2019. [26] M. Slavcheva, M. Baust, D. Cremers, and S. Ilic. Killingfusion: Non-rigid 3d reconstruction without correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1386 1395, 2017. [27] M. Slavcheva, M. Baust, and S. Ilic. Sobolevfusion: 3d reconstruction of scenes undergoing free non-rigid motion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2646 2655, 2018. [28] O. Sorkine and M. Alexa. As-rigid-as-possible surface modeling. In Symposium on Geometry processing, volume 4, pages 109 116, 2007. [29] R. W. Sumner, J. Schmid, and M. Pauly. Embedded deformation for shape manipulation. ACM Transactions on Graphics (TOG), 26(3):80 es, 2007. [30] D. Sun, X. Yang, M.-Y. Liu, and J. Kautz. Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 8934 8943, 2018. [31] C. Tang and P. Tan. Ba-net: Dense bundle adjustment network. ar Xiv preprint ar Xiv:1806.04807, 2018. [32] X. Wang, A. Jabri, and A. A. Efros. Learning correspondence from the cycle-consistency of time. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2566 2576, 2019. [33] Z. Wang, S. Li, H. Howard-Jenkins, V. Prisacariu, and M. Chen. Flownet3d++: Geometric losses for deep scene flow estimation. In The IEEE Winter Conference on Applications of Computer Vision, pages 91 98, 2020. [34] T. Yu, Z. Zheng, K. Guo, J. Zhao, Q. Dai, H. Li, G. Pons-Moll, and Y. Liu. Doublefusion: Real-time capture of human performances with inner body shapes from a single depth sensor. In Proceedings of the IEEE Conference on computer Vision and Pattern Recognition (CVPR), pages 7287 7296, 2018. [35] M. Zollhöfer, M. Nießner, S. Izadi, C. Rehmann, C. Zach, M. Fisher, C. Wu, A. Fitzgibbon, C. Loop, C. Theobalt, et al. Real-time non-rigid reconstruction using an rgb-d camera. ACM Transactions on Graphics (TOG), 33(4):1 12, 2014.