# progressive_multiview_human_mesh_recovery_with_selfsupervision__aae0e31c.pdf Progressive Multi-View Human Mesh Recovery with Self-Supervision Xuan Gong1,2, Liangchen Song1,2, Meng Zheng1, Benjamin Planche1, Terrence Chen1, Junsong Yuan2, David Doermann2, Ziyan Wu1 1 United Imaging Intelligence, Cambridge MA 02140 USA 2 University at Buffalo, Buffalo NY 14260 USA xuangong@buffalo.edu, lsong8@buffalo.edu, meng.zheng@uii-ai.com, benjamin.planche@uii-ai.com, terrence.chen@uii-ai.com, jsyuan@buffalo.edu, doermann@buffalo.edu, ziyan.wu@uii-ai.com To date, little attention has been given to multi-view 3D human mesh estimation, despite real-life applicability (e.g., motion capture, sport analysis) and robustness to single-view ambiguities. Existing solutions typically suffer from poor generalization performance to new settings, largely due to the limited diversity of image-mesh pairs in multi-view training data. To address this shortcoming, people have explored the use of synthetic images. But besides the usual impact of visual gap between rendered and target data, syntheticdata-driven multi-view estimators also suffer from overfitting to the camera viewpoint distribution sampled during training which usually differs from real-world distributions. Tackling both challenges, we propose a novel simulation-based training pipeline for multi-view human mesh recovery, which (a) relies on intermediate 2D representations which are more robust to synthetic-to-real domain gap; (b) leverages learnable calibration and triangulation to adapt to more diversified camera setups; and (c) progressively aggregates multi-view information in a canonical 3D space to remove ambiguities in 2D representations. Through extensive benchmarking, we demonstrate the superiority of the proposed solution especially for unseen in-the-wild scenarios. Introduction As a key step to several human-centric applications, 3D human mesh estimation from multi-view images has shown superiority beyond monocular image as it eliminates common ambiguities among single-image scenarios. Most successes in 3D human mesh recovery are demonstrated by supervised training. However, such models hardly generalize to in-thewild scenarios due to the lack of sufficient and diverse 3D annotations paired with multi-view images. When collected training data are insufficient and infeasible to generalize, simulated training data can be a useful alternative. Works have been done to synthetically generate and render images of human bodies for dense pose estimation (Zhu, Karlsson, and Bregler 2020), depth estimation (Varol et al. 2017), 3D pose estimation (Rogez and Schmid 2016; Varol et al. 2017; Kundu et al. 2020; Patel et al. 2021), 3D human reconstruction (Zheng et al. 2019; Copyright 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. This work was carried out during the internship of X. Gong and L. Song at United Imaging Intelligence, Cambridge, MA. Sengupta, Budvytis, and Cipolla 2020; Yu et al. 2021). Most of the self-supervised approaches only focus on single view tasks. Although works such as Kocabas, Karagoz, and Akbas (2019); Wandt et al. (2021) employ multi-view geometry for self-supervised training, for testing they infer on individual images independently, even when multi-view data are available. We believe that there commonly exist multi-view images in real scenarios, but self-supervised multi-view human pose/mesh estimation remains relatively unexplored. Some works have taken steps to eliminate the requirements of 3D human mesh annotation. Pavlakos et al. (2017) explore 3D geometry of the camera setup to lift from multiview 2D joints to 3D pictorial structure. But the training process highly relies on multi-view imagery and only estimate human pose excluding human shape. Liang and Lin (2019) generate multi-view human RGB images from existing SMPL (Loper et al. 2015) pose and shape parameters. But its synthetic paired data only helps to improve the performance of model trained with real image-3D annotations. Although these works have deliberately designed human textures, light, background during rendering, the model only trained with its synthetic data can hardly generalize to real scenarios due to the domain gap between real and synthetic images. It is obvious that the rendered images can hardly generalize to the real images in the wild. On the other hand, proxy representations (e.g., joints, silhouettes) are commonly used as intermediate representation in human mesh recovery (HMR) task, and can serve as good transition to bridge RGB image and SMPL parameters: 1) synthetic-to-real domain gap is smaller for proxy representation, thus can be more easily bridged; 2) a vast amount of SMPL parameters can be rendered into proxy representations to formulate paired synthetic data; 3) there exist multiple well-trained models predicting image to these 2D lowerdimensionality representations, which are relatively more robust and generalized compared with 3D mesh predictors. To this point, we propose to train with multi-view synthetic data for multi-view human mesh estimation. Compared to single-view synthetic training (Pavlakos et al. 2018; Sengupta, Budvytis, and Cipolla 2020; Yu et al. 2021; Gong et al. 2022; Zheng et al. 2022), two challenges specific to multi-view synthetic training arise. First, an additional domain gap between the real testing data and synthetic training data can be easily introduced due to inconsistency between The Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI-23) 2D Joint Heatmaps Multi-view Synthesis Depth Ambiguity Progressive Regression 2D-to-3D Lifting Multi-view Mesh Diversity 3D Feature Figure 1: (1) Our method is purely synthetic-data-driven regressing human mesh from 2D representations. During training 2D representations are acquired from synthetic mesh rendering. During testing 2D representations are predicted with off-the-shelf detectors. The existing well-trained 2D detectors equip our method with better generalizability and robustness to in-the-wild scenarios. (2) By unifying multi-view knowledge in canonical 3D space, we explicitly explore consensus, diversity and balance to deal with the inherent inconsistency among different views. the testing camera viewpoint distribution and the synthetic training viewpoints. Due to the lack of available multi-view SMPL parameters and camera calibration, the sampling of camera setups for multi-view synthesis should generalize to real multi-view settings which can be quite diverse across different datasets. The second challenge comes from the inherent inconsistency among multi-view representations in real scenarios. Limited by occlusion and depth ambiguity, 2D representations inferred from testing images are more likely to be biased from 3D ground-truth when compared with images which are used in those fully supervised multiview methods. To address the aforementioned issues, we propose a novel synthetic-data-driven training pipeline (Figure 1) for multiview human mesh recovery. 1) We synthetically train a regression model from multi-view 2D representations to SMPL parameters. During inference off-the-shelf detection/segmentation models are used to predict these 2D representations from RGB images. 2) The viewpoint setup for synthetic training is consistent with real testing scenarios via learnable volumetric triangulation and calibration. 3) Multiview 2D representations are aggregated in the shared 3D human space and progressively regressed to deal with the possible bias existing in 2D representation. As illustrated in Figure 2, we aim to let the regressor first focus on the consensual key area and then learn from the possible highlighted area for diversity. Once we get mesh prediction from these two regression iterations, we are able to better balance the multi-view volumetric features via reprojection consistency where views less consistent with the predicted mesh will be given less weight in the final mesh refinement iteration. Empirical evaluations show that our method achieves very competitive results on H3.6M (Ionescu et al. 2013), Total Capture (Trumble et al. 2017) and challenging Ski Pose (Sp orri 2016; Rhodin et al. 2018) dataset compared with other fully/weakly supervised multi-view human mesh recovery and human pose estimation methods. Our key contributions can be summarized as: 1) We propose a multi-view synthetic-data-driven training pipeline for multi-view human mesh recovery, mapping multi-view 2D representations to shared 3D human space to bridge the real-synthetic gap. 2) We progressively regress multi-view representations by first exploring the consensus and diversity among views in 3D space and then reaching evidential balance among views. This design can efficiently tolerate the bias commonly existing in single 2D representation thus generalizable to in-the-wild scenarios. 3) We conduct extensive experiments on standard benchmark datasets and demonstrate comparable numbers with fully/weakly supervised methods on conventional evaluation metrics. Related Works Monocular 3D Human Pose Estimation 3D human pose estimation (HPE) (Agarwal and Triggs 2005; Song et al. 2021) problem can be categorized into 3D body keypoint/skeleton prediction and 3D human mesh recovery, based on representing the human body with kinematic or volumetric models. On the other hand, 3D human mesh recovery (HMR) (Loper et al. 2015) regresses and outputs mesh parameters, containing richer shape and texture information of the human body. Recently, numerous methods (Kolotouros et al. 2019; Arnab, Doersch, and Zisserman 2019; Bogo et al. 2016; Li et al. 2021) focus on estimating parameters of the Skinned Multi-Person Linear (SMPL) (Loper et al. 2015), a commonly-used volumetric human model with high compatibility, to statistically regress human meshes. Several works take steps to leverage a variety of easily-obtained clues, i.e., weak supervision, such as paired 2D landmarks and silhouettes (Tan, Budvytis, and Cipolla 2017; Pavlakos et al. 2018; Kanazawa et al. 2018; Rong et al. 2019; Wehrbein et al. 2021). Multi-View 3D Human Pose Estimation Many methods (Dong et al. 2019; Liang and Lin 2019; Qiu et al. 2019; Rhodin, Salzmann, and Fua 2018; Pavlakos et al. 2017; Zhang et al. 2021) have recently proposed for multiview 3D HPE. While the majority focuses on 3D body keypoint/skeleton prediction (Pavlakos et al. 2017; Dong et al. 2019; Rhodin, Salzmann, and Fua 2018), we consider the problem of multi-view 3D HMR, which reconstructs 3D SMPL (Loper et al. 2015) pose and shape parameters given multiple view images. Existing multi-view HMR methods are all supervised, fusing multi-view with probabilistic modeling (Kolotouros et al. 2021) or collaborative learning (Li, Oskarsson, and Heyden 2021) to regress SMPL parameters. Liang and Lin (2019) uses additional synthetic image-SMPL pairs to train a multi-view multi-stage regression network. (Dong et al. 2021) aggregates multi-view observation based on confidence-aware majority voting technique. Method Prerequisites 3D human mesh parameterization. Skinned Multi Person Linear (SMPL) (Loper et al. 2015) is a parametric model providing independent body shape β and pose θ parameters with very low-dimensional parameters (i.e., β R10 and θ R72). Pose parameters θ = {θg, θj} include global body rotation θg (3-DOF) and relative 3D rotations of 23 joints θj (23 3-DOF) in the axis-angle format. Shape parameters include individual heights and weights indicated by the first 10 coefficients of a PCA shape space. SMPL provides a differentiable kinematic function S from {θ, β} to 6,890 mesh vertices: v = S(θ, β) R6890 3. Besides, 3D locations for NJ joints of interest are obtained as j3D = J v, where J RNJ 6890 is a linear regression matrix. Monocular training data synthesis. Existing synthetic based HMR methods (Sengupta, Budvytis, and Cipolla 2020, 2021b,a) generate paired 2D representations (i.e., binary mask, edge and 2D joints) and 3D meshes with SMPL parameters on the fly during training process. At each training step, pose {θg, θj} and shape β are sampled from Mo Cap (C 2003; Rogez and Schmid 2016) datasets and prior statistical normal distribution respectively. The camera translation T R3 is also dynamically sampled from prior distribution. The intrinsic parameters are fixed and represented by focal length f R2 and image center offset t = [H/2, W/2], where H, W is rendered image size. {θg, θj, β} are forwarded into the SMPL model to obtain 3D joints j3D. 2D joints j2D can be acquired by j2D = fΠ(j3D + T ) + t, where Π denotes perspective projection. The 2D joints j2D RNJ 2 are transformed into 2D Gaussian joint heatmaps J RNJ H W . Another 2D proxy representation, human mask, can be represented by M RH W . The training model utilizes the synthesized paired data with 2D representations {J, M} as input and 3D meshes {θ, β, v, j3D} as output. Volumetric triangulation. Beyond basic algebraic triangulation, volumetric triangulation (Iskakov et al. 2019) is able to unproject multi-view 2D features along projection rays to fill a shared 3D cube. The cube is a L L L - sized 3D bounding box in the global space discretized by G G G volumetric grids, where G represents the number of voxels along each axis. Then each voxel is filled with the global coordinates of the voxel center to get V coords RG G G 3. V coords is projected to the image plan to get its corresponding 2D pixel index V proj RG G G 2. Given 2D maps F RC H W in image space, we can fill a cube V RG G G C by bilinear sampling (Jaderberg et al. 2015) using V proj. The whole process is differentiable and agnostic to the number of views. Multi-View Training Data Synthesis As described above, the synthesis of 2D proxy representations relies on camera extrinsic parameters {Rh c, Th c} (the transformation between human coordinate and camera coordinate) for each camera. The intuitive solution of multiview synthesis is to randomly sample multi-view camera extrinsic parameters around the human body. But the lack of statistical priors makes it not ideal since some views (e.g., from below the human body) can never happen in real scenarios. It also requires manual tuning to ensure as much visible space as possible to avoid large area of blind spot. We first sample one camera setting {Rh c1, Th c1} from prior and then extend it to N cameras according to the transformations among cameras in real scenarios. Though there exists no direct information, w.r.t., relations among camera setups, it can be inferred from the public camera calibrations for multi-view datasets (Ionescu et al. 2013; Trumble et al. 2017). Given {Rw cn, Tw cn}(n = 1...N) as the rotation and translation from canonical world coordinate to the n-th camera coordinate, we calculate the transformation from the first camera to the other cameras: Rc1 cn = Rw cn (Rw c1)T, Tc1 cn = Tw cn Rc1 cn Tw c1 (n = 2...N). (1) Utilizing these priors ensures that our synthetically trained model can better generalize to different testing multi-view images in the wild. The overall multi-view synthesis process at each training step can be summarized as: (1) sample {θg, θj, β, T } following the same strategy as Sengupta, Budvytis, and Cipolla (2020); (2) sample one set of camera settings {Rw cn, Tw cn|n = 1...N} (e.g., N from 2 to 8); (3) calculate the transformation from camera 1 to the other cameras using Eq.1; (4) regress and render 2D joints heatmaps and binary mask {J1, M 1} under camera-1 using Rh c1 (from θg) and Th c1 as camera extrinsic parameters; (5) regress and render 2D representations {Jn, M n} under other cameras using camera extrinsic parameters below: Rh cn = Rc1 cn Rh c1, Th cn = Rh cn Th c1 + Tc1 cn (n = 2...N). (2) We forward {0, θj, β} into the SMPL statistical model and get {v, j3D} in human coordinate system. We further transform {v, j3D} with individual camera rotation Rh cn to {vn, jn 3D} making it specific under each camera rotation. Then we project to 2D joints {Jn and render to get binary mask M n} for each camera n. The training model utilizes the synthesized paired data with 2D representations {Jn, M n} and 3D meshes {θj, β, vn, jn 3D} where θj and β are shared across all views. Note that the aforementioned camera extrinsic {Rw cn, Tw cn} and intrinsic matrix K will also be used in forwarding. S1: Volumetric Feature Sampling with Consensus S2: Volumetric Feature Sampling with Diversity S3: Multi-view Weighted Fusion with Per-View Consistency Feedback Learnable Multi-View Volumetric Feature Triangulation Per-View Human-Camera Projection Parameters Learnable calibration and transformation Multi-View Progressive Mesh Regression Per-View 2D Feature Map Multi-View Rendering Sampled from Prior Volumetric Projection Lifted Per-View Feature Map Ɵ3 Per-View Weights Consistency Maps Mesh Regressor Union Masking Intersection Masking Figure 2: Overall illustration of the proposed pipeline. We unproject the 2D occupancy maps to 3D space to obtain intersection and union of 3D occupancy used to mask the volumetric features. The intersection and union masking in 3D space help to sample volumetric features according to multi-view consensus and diversity respectively. This progressive regression design can efficiently tolerate possible bias and seek consistency in real multi-view settings. Learnable Volumetric Calibration The 3D space for fusion is designed to be consistent with the human coordinate rather than world coordinate for mesh regression efficiency and model generalizability. However, for real testing images, we have no access to the transformation from human coordinate to image space which is necessary to do the volumetric triangulation. On the other hand, the camera intrinsic parameters K are always known and some testing datasets provide transformation from canonical world coordinate to camera coordinate. When this transformation is not available, we can simply define a world coordinate initialized with camera-1. The transformation between camera-1 and other cameras can be acquired from 2D joints in different views. Please refer to Kocabas, Karagoz, and Akbas (2019) for details. Based on the aforementioned known transformation from canonical world to each camera coordinate (Rw cn, Tw cn), we design volumetric calibration to transform from canonical world to human coordinate ( ˆRw h, ˆTw h) so that multi-view 2D space can be unified in a shared human space. Initialized with camera-1, the canonical world is first translated to be with the same origin as the human center which can be interpolated w.r.t. statistical torso length and 2D joints, and then rotated to be consistent with human global pose. Translation estimation. Under each camera view, we obtain the 3D representation of pelvis heatmaps via volumetric triangulation (Iskakov et al. 2019). The per-voxel likelihood for pelvis is obtained by summing up multi-view 3D heatmaps. Via argmax (averaging the 3D positions of the voxels if there are multiple voxels containing the maximum value) we estimate the 3D pelvis position in world space as the translation from world origin to human origin ˆTw h,w under world space: ˆTw h,w = argmax X n V(Jn pelvis; Rw cn, Tw cn, K), (3) where V( ; ) represents volumetric triangulation. Note we indicate all transformation by uniform sequence with rotation first and then translation, we therefore organize the transformation from human to each camera coordinate: ˆRh cn = Rw cn ˆRT w h, ˆTh cn = Tw cn ˆRw cn ˆTw h,w. (4) Rotation learning. For each camera view n, the encoder takes 2D joints heatmaps and binary mask {Jn, M n} as input and output downsized features F n RC H0 W0, where C is channel size. To learn the rotation from canonical world to human, we unproject F n to the corrected 3D space with human origin and world rotation (acquired by the estimated translation mentioned above). We first take average of the unprojected volumetric 3D features among views, then forward the flattened volumetric features into a fullyconnected layer FR: ˆRw h = FR( 1 n V(F n; ˆRw cn, ˆTh cn, K, H (5) where H H0 , W W0 are the scale factors from image space to its downsampled feature space. The output of the fullyconnected layer is a continuous 6-dimensional representa- tion (Zhou et al. 2019) which can be converted to a discontinuous Euler rotation matrix ˆRw h. Note that both volumetric triangulation (grid sampling) and transformation are differentiable, thus ˆRw h can be learned via projecting the predicted 3D mesh to 2D. We note that the aforementioned calibration from canonical world to human is to unify diverse camera setups in common human space for efficient multi-view learning. The currently available K and ˆTh cn is not learnable and only a rough estimation which aims to correct the 3D cube center to be around the pelvis center. For more accurate joint learning together with the predicted 3D mesh, we employ reprojection loss with orthographic projection following Kanazawa et al. (2018). Specifically, we learn camera parameters from a fully-connected layer FOP for individual view: ˆπn = FOP(F n), where ˆπn = [ˆπn s , ˆπn t ], ˆπn s R is the scale factor and ˆπn t R2 is translation. Progressive Multi-View Aggregation From the aforementioned volumetric calibration, we are able to obtain the volumetric features V n F RG G G C for each camera-n in the uniform 3D human space: V n F = V(F n; ˆRh cn, ˆTh cn, K, H V n F from multiple views are fused (averaged or weighted summation), flattened, and then passed to the regressor R to predict pose and shape parameters Θ = { ˆθj, ˆβ}. Note that we only predict 23 joints rotation ˆθj shared by all views (excluding global root orientation) as we have learnt viewspecific global orientation ˆRh cn in volumetric calibration. Following the standard iterative error feedback (IEF) procedure (Kanazawa et al. 2018), we employ three iterative regressions to optimize Θ and further propose progressive multi-view aggregation. We introduce how we progressively learn from consensus/diversity information from intersection and union occupancy mask, and then how we balance the multi-view 3D features from the consistency between consensus and diversity preserving mesh prediction. Consensus and diversity sampling. Under each cameran, we consider all the possible nonzero areas of 2D joints heatmaps Jn and binary mask M n as 2D occupancy mask On {0, 1}H W and then we obtain the volumetric occupancy mask in 3D human space V n O {0, 1}G G G: On = 1(( XNj j=1 Jn j + M n) > 0), V n O = V(On; ˆRh cn, ˆTh cn, K). (7) To efficiently aggregate the volumetric occupancy masks, we take intersection of {V n O |n = 1...N} as V I O representing the area of interest shared by all views (consensus), and take union of them as V U O which representing the area of interest masked by at least one view (diversity): V I O = minn V n O , V U O = maxn V n O , (8) where V I O, V U O {0, 1}G G G. To achieve consensus while also maintaining the diversity that is inherent among the multiple views, we mask the 3D volumetric features {V n F } spatial-wisely with occupancy intersection V I O and occupancy union V U O respectively: V I F = V I O 1 n V n F , V U F = V U O 1 n V n F . (9) where is Hadamard product, and V I F, V U F RG G G C indicate the fused 3D features in consensus and diversity occupancy area respectively. Note that the diversity occupancy area is introduced to tolerate possible bias of one-view 3D occupancy which is easily caused by inaccurate camera calibration. We progressively forward V I F, V U F into the regressor R so that the regressor can first focus on the features in area commonly occupant by all views (V I O), and then consider the features in all possible occupant areas (V U O ). The output of R can be represented by Θ1 and Θ2: Θ1 = Θ0 + R(V I F; Θ0), Θ2 = Θ1 + R(V U F ; Θ1), (10) where Θ0 is reposed pose and mean shape for initialization. Multi-view balance via consistency weighting. We take average of the multi-view 3D features for intersection and union fusion in the first two regression iteration. At the last iteration of regression we utilize the current 3D mesh prediction as evidence to seek consistency among views. Specifically, we project the 3D mesh to individual 2D image spaces for spatial-wise consistency as fusion confidence under each camera. Given the output of the regressor after the second iteration Θ2 = {ˆθj, ˆβ}, SMPL takes {0, ˆθj, ˆβ} to infer 3D vertices ˆv and 3D joints ˆj3D. Using a differentiable renderer (Ravi et al. 2020), we generate a body mask ˆ M n from ˆv according to the camera parameters { ˆRh cn, ˆπn}. We also obtain the reprojected 2D joints ˆjn 2D = ˆπn s Π( ˆRh cn ˆj T 3D)+ ˆπn t and convert to heatmap version ˆ Jn for each camera-n. Comparing these reprojected 2D representations { ˆ Jn, ˆ M n} with the input {Jn, M n}, we calculate the consistency map ϕn RH W under each 2D image space: ϵ + | ˆ M n M n| + PNj j=1 | ˆ Jn j Jn j |/Nj . (11) We further unproject ϕn to get the volumetric consistency representation V n ϕ RG G G of each view under the commonly shared 3D human space, and then normalize these volumetric consistency among N camera views: V n ϕ = V n ϕ P n V n ϕ = V(ϕn; ˆRh cn, ˆTh cn, K) P n V(ϕn; ˆRh cn, ˆTh cn, K) . (12) Taking V n ϕ as view-specific volumetric confidence, we are able to balance volumetric 3D features V n F under each camera into V B F where the view less consistent with jointly reached 3D mesh is given less confidence for fusion: n V n ϕ V n F . (13) At the final iteration, the regressor takes the consistency balanced 3D feature V B F for the final prediction :Θ3 = Θ2 + R(V B F ; Θ2). Method Training Requirement Metrics Multi-View Imagery Temporal Sequence Rhodin et al. (2018) J3D 131.7 98.2 PVH-TSP (2017) J3D 87.3 - Tome et al. (2018) J3D 52.8 - Remelli et al. (2020) J3D 30.2 - Bartol et al. (2022) J3D 29.1 - Pavlakos et al. (2017) 56.9 - Trumble et al. (2018) J3D 62.5 - Liang et al. (2019) Mesh 79.8 45.1 Li et al. (2021) Mesh 64.8 43.8 Pro HMR (2021) Mesh 62.2 34.5 Ours 53.8 42.4 Table 1: Comparisons on MPJPE and PMPJPE (both in mm) on the Human3.6M test sets with multi-view human pose/mesh estimation methods. indicates method which can recover human shape beyond human pose. Loss Function As described above, from the final description Θ3 = {ˆθj, ˆβ} we have vertices ˆv and 3D joints ˆj3D based on SMPL regression. With camera parameters { ˆRh cn, ˆπn}, we infer to vertices and 3D joints under each camera rotation: ˆvn = ˆRh cn ˆv T, ˆjn 3D = ˆRh cn ˆj T 3D. ˆjn 3D is then projected to 2D joints ˆjn 2D with orthographic projection. The overall loss for mesh regression is therefore defined as L( ˆRh cn, ˆθj, ˆβ, ˆvn, ˆjn 2D, ˆjn 3D, Rh cn, θj, β, , vn, jn 2D, jn 3D) =ωθL2(ˆθj, θj) + ωβL2( ˆβ, β) + n=1 ωRL2( ˆRh cn, Rh cn) ωv L2(ˆvn, vn) + ωj2DL2(ˆjn 2D, jn 2D) + ωj3DL2(ˆjn 3D, jn 3D) (14) where L2 denotes the mean square error (MSE), and ωθ, ωβ, ωR ωv, ωj2D, and ωj3D indicate weights for joints pose and shape, view-specific global pose, vertices, 2D joints, 3D joints respectively. Note for L2(ˆθj, θj) both are first converted to rotation matrix for the MSE loss. Experiments Note that we train only one model generalized to different test sets, whereas most counterparts train individual model, with corresponding training data, for each specific test set. Datasets and Metrics Training data. Following the protocol by Sengupta, Budvytis, and Cipolla (2020), we sample SMPL pose parameters from the training sets of UP-3D (Lassner et al. 2017), 3DPW (von Marcard et al. 2018), and the five training subjects (S1, S5, S6, S7, S8) of Human3.6M (Ionescu et al. 2013). Evaluation data. To evaluate the generalizability of our method, we test on both indoor and outdoor datasets with different number of views. Human3.6M is one of the largest/most commonly used 3D human pose estimation benchmark with SMPL annotation. Besides, we evaluate on Total Capture (in-the-studio) and Ski Pose (in-the-wild). Both only have 3D joint annotations (no body shape ones). Human3.6M: The Human3.6M dataset (Ionescu et al. 2013) provides a total of 3.6 million frames in synchronized four-views. The camera placement is slightly different for each of the seven subjects. We follow the most popular protocol 1, testing on subjects S9, S11. We report mean per joint position error (MPJPE) and mean per joint position error after rigid alignment with Procrustes analysis (PMPJPE) on the 17 joints in H3.6M definition. Total Capture: Total Capture dataset (Trumble et al. 2017) consists of 1.9 million frames, captured from 8 calibrated full HD video cameras recording at 60Hz. Following the typical data split (Trumble et al. 2017), we use Walking2 , Freestyle3 , and Acting3 on subjects 1, 2, 3, 4, 5 for testing. We report the mean per joint position error (MPJPE) as 3D pose metric for comparison with prior arts. Ski-Pose PTZ: This dataset (Sp orri 2016; Rhodin et al. 2018) records competitive alpine skiers performing giant slalom runs with eight moving cameras. The cameras are rotating and zooming to keep the alpine skier in the field of view. We follow the typically used metrics: MPJPE, PMPJPE, and percentage of correct keypoints (PCK) thresholded at 150mm (Mehta et al. 2017). Implementation Details Synthetic training. We generate multi-view paired data on the fly. To simulate noise and discrepancy between 2D joints and mask prediction and among different views, we apply a series of processing and augmentations. Training is done using Adam (Kingma and Ba 2014) optimizer for 6 epochs with a learning rate of 1e 4 and a batch size of 16. It takes ~3 days on one A100 GPU. Testing. We infer 2D joints on the testing images with the pretrained Keypoint-RCNN (He et al. 2017) with Res Net50 backbone. We predict the human mask using pretrained Dense Pose-RCNN (He et al. 2017) with Res Net-101 backbone. For consistency with training, we crop both the masks and 2D joints heatmaps with a scale of 1.2 before forwarding them to the network for 3D mesh inference . Results Human3.6M. Table 1 compares our method on test sets of Human3.6M Protocol 1 with other multi-view human pose estimation methods. The counterparts utilize fully-paired 3D annotation or auxiliary clues, e.g., multi-view images. Compared with the other self-supervised method (Pavlakos et al. 2017), our method 1) is able to predict human shape beyond human pose, 2) does not rely on any auxiliary requirement. Comparing to the pose-only methods (top half), we note a large performance gap between self-supervised methods and fully-supervised arts. But our self-supervised mesh recovery method is comparable to the fully-supervised mesh recovery SOTA method (Kolotouros et al. 2021). Total Capture. Table 2 compares our HMR method with other multi-view human pose/mesh estimation methods on Total Capture test set. All the methods for comparison take Superv. Method Subjects (S1,2,3) Subjects (S4,5) Mean Walking2 Acting3 Freestyle3 Walking2 Acting3 Freestyle3 Image-Joints3D pairs (In) PVH (2017) 48.3 94.3 122.3 84.3 154.5 168.5 107.3 Tri-CPM (2016) 79.0 106.5 112.1 79.0 73.7 149.3 99.8 IMUPVH (2017) 30.0 49.0 90.6 36.0 109.2 112.1 70.9 Image-Joints3D pairs (In) Trumble et al. (2018) 42.0 59.8 120.5 58.4 103.4 162.1 85.4 Image-Mesh pairs (Cr) Pro HMR (2021) 125.7 118.9 134.3 131.9 125.2 135.8 127.8 Self Ours 66.1 69.3 58.9 64.4 79.1 61.3 64.2 Table 2: Comparison of multi-view 3D human pose/mesh estimation methods in terms of 3D pose errors MPJPE (mm) on Total Capture test set. indicates method which can recover both body pose and shape. In denotes training data from the same dataset as the testing data. Cr denotes training data is from cross/different dataset. We report the results of Pro HMR trained with Human3.6M and additional 2D keypoints fitting. Method Train Test Pavllo et al. (2019) T T 106.0 88.1 - Adapt Pose(2021) T S 99.4 83.0 - Rhodin et al. (2018) MV S 85.0 - - Canon Pose (2021) MV S 128.1 89.6 67.1 Pro HMR (2021) S(Cr) S 122.7 82.6 73.4 Pro HMR (2021) S(Cr) MV 105.7 73.1 80.3 Ours S 109.2 72.6 77.5 Ours MV 89.6 64.8 86.0 Table 3: Comparison with 3D pose/mesh estimation methods on Ski-Pose PTZ test set in terms of MPJPE, PMPJPE (in mm) and PCK (%). indicates additional ability to recover body shape. S , T , MV denotes single, temporal consistent, and multi-view images respectively. Cr denotes cross dataset (Human3.6M) different with the test data. all view images as input for inference. We note that our self-supervised human pose and shape estimation method demonstrates better performance in terms of MPJPE (mm) than the fully-supervised methods (Trumble et al. 2017; Wei et al. 2016; Trumble et al. 2018) which are only able to predict human pose. The superiority of our method beyond the others is more obvious on S4,5 which are unseen subjects in Total Capture training set, as those methods trained with paired image-3D annotation can hardly generalize. We can see the SOTA multi-view mesh predication method (Kolotouros et al. 2021) (trained with Human3.6M) can hardly generalize to Total Capture test set though they are both inthe-studio datasets. In contrast, our model trained with synthetic 2D-3D pairs better generalize to diverse unseen data. Ski Pose PTZ. The comparison on the test set of Ski Pose is shown in Table 3. With additional utility of body shape estimation, our method still outperforms self-supervised SOTA human pose estimation method (Wandt et al. 2021). Comparison with SOTA supervised multi-view human mesh recovery method (Pro HMR (Kolotouros et al. 2021) trained with Human3.6M) shows our purely synthetic-data-trained method has superior generalization to in-the-wild scenarios even when Pro HMR uses 2D keypoints fitting during testing while ours only use 2D keypoints as input. Qualitative results. Figure 3 gives qualitative examples Figure 3: Visualization of reconstructed SMPL mesh for Human3.6M (row 1), Ski Pose (rows 2) and Total Capture (row 3) datasets (select four views for latter two datasets). where we visualize our predicted 3D mesh on images from the three testing datasets. Notably, we train only one model and test on different datasets. The results demonstrate the robustness and generalization ability of our syntheticallytrained model to various unseen in-the-wild data. To the best of our knowledge, we propose the first multiview human mesh recovery method based on self-supervised synthetic training. Our solution first extracts intermediate 2D representations from each view and projects the corresponding features to a 3D canonical space with learnable volumetric calibration. Multi-stage progressive regressors then iteratively refine estimated mesh parameters based on different feature-sampling criteria. Extensive evaluations demonstrate the efficacy and superior performance of the proposed method, especially in challenging in-the-wild scenarios where 1) single-view-based methods suffer from depth ambiguities and 2) supervision-based methods have no access to any prior of in-domain image and annotation. References Agarwal, A.; and Triggs, B. 2005. Recovering 3D human pose from monocular images. IEEE transactions on pattern analysis and machine intelligence, 28(1): 44 58. Arnab, A.; Doersch, C.; and Zisserman, A. 2019. Exploiting temporal context for 3D human pose estimation in the wild. In CVPR. Bartol, K.; Bojani c, D.; Petkovi c, T.; and Pribani c, T. 2022. Generalizable Human Pose Triangulation. In CVPR, 11028 11037. Bogo, F.; Kanazawa, A.; Lassner, C.; Gehler, P.; Romero, J.; and Black, M. J. 2016. Keep it SMPL: Automatic estimation of 3D human pose and shape from a single image. In ECCV, 561 578. Springer. C. 2003. Mo Cap. In mocap. cs. cmu. Dong, J.; Jiang, W.; Huang, Q.; Bao, H.; and Zhou, X. 2019. Fast and Robust Multi-Person 3D Pose Estimation from Multiple Views. Dong, Z.; Song, J.; Chen, X.; Guo, C.; and Hilliges, O. 2021. Shape-aware multi-person pose estimation from multi-view images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 11158 11168. Gholami, M.; Wandt, B.; Rhodin, H.; Ward, R.; and Wang, Z. J. 2021. Adapt Pose: Cross-Dataset Adaptation for 3D Human Pose Estimation by Learnable Motion Generation. ar Xiv preprint ar Xiv:2112.11593. Gong, X.; Zheng, M.; Planche, B.; Karanam, S.; Chen, T.; Doermann, D.; and Wu, Z. 2022. Self-supervised Human Mesh Recovery with Cross-Representation Alignment. In European Conference on Computer Vision, 212 230. Springer. He, K.; Gkioxari, G.; Doll ar, P.; and Girshick, R. 2017. Mask r-cnn. In ICCV, 2961 2969. Ionescu, C.; Papava, D.; Olaru, V.; and Sminchisescu, C. 2013. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. PAMI, 36(7): 1325 1339. Iskakov, K.; Burkov, E.; Lempitsky, V.; and Malkov, Y. 2019. Learnable triangulation of human pose. In ICCV, 7718 7727. Jaderberg, M.; Simonyan, K.; Zisserman, A.; et al. 2015. Spatial transformer networks. Neur IPS, 28. Kanazawa, A.; Black, M. J.; Jacobs, D. W.; and Malik, J. 2018. End-to-end recovery of human shape and pose. In CVPR, 7122 7131. Kingma, D. P.; and Ba, J. 2014. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980. Kocabas, M.; Karagoz, S.; and Akbas, E. 2019. Selfsupervised learning of 3d human pose using multi-view geometry. In CVPR, 1077 1086. Kolotouros, N.; Pavlakos, G.; Black, M. J.; and Daniilidis, K. 2019. Learning to reconstruct 3D human pose and shape via model-fitting in the loop. In ICCV, 2252 2261. Kolotouros, N.; Pavlakos, G.; Jayaraman, D.; and Daniilidis, K. 2021. Probabilistic Modeling for Human Mesh Recovery. In ICCV. Kundu, J. N.; Seth, S.; Jampani, V.; Rakesh, M.; Babu, R. V.; and Chakraborty, A. 2020. Self-supervised 3D human pose estimation via part guided novel image synthesis. In CVPR, 6152 6162. Lassner, C.; Romero, J.; Kiefel, M.; Bogo, F.; Black, M. J.; and Gehler, P. V. 2017. Unite the people: Closing the loop between 3d and 2d human representations. In CVPR, 6050 6059. Li, J.; Xu, C.; Chen, Z.; Bian, S.; Yang, L.; and Lu, C. 2021. Hybr IK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation. In CVPR. Li, Z.; Oskarsson, M.; and Heyden, A. 2021. 3D Human Pose and Shape Estimation Through Collaborative Learning and Multi-View Model-Fitting. In WACV, 1888 1897. Liang, J.; and Lin, M. C. 2019. Shape-aware human pose and shape reconstruction using multi-view images. In ICCV, 4352 4362. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; and Black, M. J. 2015. SMPL: A skinned multi-person linear model. ACM transactions on graphics (TOG), 34(6): 1 16. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; and Theobalt, C. 2017. Monocular 3d human pose estimation in the wild using improved cnn supervision. In 2017 international conference on 3D vision (3DV), 506 516. IEEE. Patel, P.; Huang, C.-H. P.; Tesch, J.; Hoffmann, D. T.; Tripathi, S.; and Black, M. J. 2021. AGORA: Avatars in geography optimized for regression analysis. In CVPR, 13468 13478. Pavlakos, G.; Zhou, X.; Derpanis, K. G.; and Daniilidis, K. 2017. Harvesting multiple views for marker-less 3d human pose annotations. In CVPR, 6988 6997. Pavlakos, G.; Zhu, L.; Zhou, X.; and Daniilidis, K. 2018. Learning to estimate 3D human pose and shape from a single color image. In CVPR, 459 468. Pavllo, D.; Feichtenhofer, C.; Grangier, D.; and Auli, M. 2019. 3d human pose estimation in video with temporal convolutions and semi-supervised training. In CVPR, 7753 7762. Qiu, H.; Wang, C.; Wang, J.; Wang, N.; and Zeng, W. 2019. Cross View Fusion for 3D Human Pose Estimation. In ICCV. Ravi, N.; Reizenstein, J.; Novotny, D.; Gordon, T.; Lo, W.- Y.; Johnson, J.; and Gkioxari, G. 2020. Accelerating 3D Deep Learning with Py Torch3D. ar Xiv:2007.08501. Remelli, E.; Han, S.; Honari, S.; Fua, P.; and Wang, R. 2020. Lightweight multi-view 3d pose estimation through cameradisentangled representation. In CVPR, 6040 6049. Rhodin, H.; Salzmann, M.; and Fua, P. 2018. Unsupervised geometry-aware representation for 3d human pose estimation. In ECCV, 750 767. Rhodin, H.; Sp orri, J.; Katircioglu, I.; Constantin, V.; Meyer, F.; M uller, E.; Salzmann, M.; and Fua, P. 2018. Learning monocular 3d human pose estimation from multi-view images. In CVPR, 8437 8446. Rogez, G.; and Schmid, C. 2016. Mo Cap-guided Data Augmentation for 3D Pose Estimation in the Wild. In Neur IPS, 3108 3116. Rong, Y.; Liu, Z.; Li, C.; Cao, K.; and Loy, C. C. 2019. Delving deep into hybrid annotations for 3d human recovery in the wild. In ICCV, 5340 5348. Sengupta, A.; Budvytis, I.; and Cipolla, R. 2020. Synthetic training for accurate 3d human pose and shape estimation in the wild. In BMVC. Sengupta, A.; Budvytis, I.; and Cipolla, R. 2021a. Hierarchical Kinematic Probability Distributions for 3D Human Shape and Pose Estimation From Images in the Wild. In ICCV, 11219 11229. Sengupta, A.; Budvytis, I.; and Cipolla, R. 2021b. Probabilistic 3D Human Shape and Pose Estimation From Multiple Unconstrained Images in the Wild. In CVPR, 16094 16104. Song, L.; Yu, G.; Yuan, J.; and Liu, Z. 2021. Human pose estimation and its application to action recognition: A survey. Journal of Visual Communication and Image Representation, 76: 103055. Sp orri, J. 2016. Reasearch dedicated to sports injury prevention-the sequence of prevention on the example of alpine ski racing. Habilitation with Venia Docendi in Biomechanics, 1(2): 7. Tan, J.; Budvytis, I.; and Cipolla, R. 2017. Indirect deep structured learning for 3D human body shape and pose prediction. In BMVC. Tome, D.; Toso, M.; Agapito, L.; and Russell, C. 2018. Rethinking pose in 3d: Multi-stage refinement and recovery for markerless motion capture. In International Conference on 3D vision (3DV), 474 483. IEEE. Trumble, M.; Gilbert, A.; Hilton, A.; and Collomosse, J. 2018. Deep autoencoder for combined human pose estimation and body model upscaling. In ECCV, 784 800. Trumble, M.; Gilbert, A.; Malleson, C.; Hilton, A.; and Collomosse, J. 2017. Total Capture: 3D Human Pose Estimation Fusing Video and Inertial Sensors. In BMVC. Varol, G.; Romero, J.; Martin, X.; Mahmood, N.; Black, M. J.; Laptev, I.; and Schmid, C. 2017. Learning from synthetic humans. In CVPR, 109 117. von Marcard, T.; Henschel, R.; Black, M. J.; Rosenhahn, B.; and Pons-Moll, G. 2018. Recovering accurate 3d human pose in the wild using imus and a moving camera. In ECCV, 601 617. Wandt, B.; Rudolph, M.; Zell, P.; Rhodin, H.; and Rosenhahn, B. 2021. Canon Pose: Self-supervised monocular 3D human pose estimation in the wild. In CVPR, 13294 13304. Wehrbein, T.; Rudolph, M.; Rosenhahn, B.; and Wandt, B. 2021. Probabilistic Monocular 3D Human Pose Estimation With Normalizing Flows. In ICCV, 11199 11208. Wei, S.-E.; Ramakrishna, V.; Kanade, T.; and Sheikh, Y. 2016. Convolutional pose machines. In CVPR, 4724 4732. Yu, Z.; Wang, J.; Xu, J.; Ni, B.; Zhao, C.; Wang, M.; and Zhang, W. 2021. Skeleton2Mesh: Kinematics Prior Injected Unsupervised Human Mesh Recovery. In ICCV, 8619 8629. Zhang, Y.; Li, Z.; An, L.; Li, M.; Yu, T.; and Liu, Y. 2021. Lightweight multi-person total motion capture using sparse multi-view cameras. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 5560 5569. Zheng, M.; Planche, B.; Gong, X.; Yang, F.; Chen, T.; and Wu, Z. 2022. Self-supervised 3d patient modeling with multi-modal attentive fusion. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 115 125. Springer. Zheng, Z.; Yu, T.; Wei, Y.; Dai, Q.; and Liu, Y. 2019. Deephuman: 3d human reconstruction from a single image. In ICCV, 7739 7749. Zhou, Y.; Barnes, C.; Lu, J.; Yang, J.; and Li, H. 2019. On the continuity of rotation representations in neural networks. In CVPR, 5745 5753. Zhu, T.; Karlsson, P.; and Bregler, C. 2020. Simpose: Effectively learning densepose and surface normals of people from simulated data. In ECCV, 225 242. Springer.