# synthetictoreal_pose_estimation_with_geometric_reconstruction__bb81dc3b.pdf Synthetic-to-Real Pose Estimation with Geometric Reconstruction Qiuxia Lin1 Kerui Gu1 Linlin Yang2,3 Angela Yao1 1Department of Computer Science, National University of Singapore 2State Key Laboratory of Media Convergence and Communication, CUC 3School of Information and Communication Engineering, CUC {qiuxia, keruigu, ayao}@comp.nus.edu.sg Pose estimation is remarkably successful under supervised learning, but obtaining annotations, especially for new deployments, is costly and time-consuming. This work tackles adapting models trained on synthetic data to real-world target domains with only unlabelled data. A common approach is model fine-tuning with pseudo-labels from the target domain; yet many pseudo-labelling strategies cannot provide sufficient high-quality pose labels. This work proposes a reconstructionbased strategy as a complement to pseudo-labelling for synthetic-to-real domain adaptation. We generate the driving image by geometrically transforming a base image according to the predicted keypoints and enforce a reconstruction loss to refine the predictions. It provides a novel solution to effectively correct confident yet inaccurate keypoint locations through image reconstruction in domain adaptation. Our approach outperforms the previous state-of-the-arts by 8% for PCK on four large-scale hand and human real-world datasets. In particular, we excel on endpoints such as fingertips and head, with 7.2% and 29.9% improvements in PCK. 1 Introduction Training deep models for human pose estimation, be it of the full body [30, 29, 2] or hands [37, 42], requires large-scale labelled datasets. Synthetic data [43, 31] has become more favoured due to the ease of generation and the challenges of accurately annotating real-world data. Despite advancements in realistic rendering [23, 5], models trained on synthetic data do not generalize well to realworld settings. The gap arises due to variations in pose, appearance, lighting, and other low-level differences between the two data domains. This paper aims to close this gap and targets learning from labelled synthetic and unlabelled real-world data for application to real-world settings. Domain adaptation studies the transfer of models from synthetic to real data. Recent works on domain adaptive pose estimation [17, 36, 15] learn from target pseudo-labels. However, the gains from pseudo-labelling can be limited. There is an inherent contradiction only labels which are sufficiently accurate benefit the learning, and it is important to ensure label quality through selection [15] and or correction schemes [36, 18]. Yet the ability to obtain high-quality pseudo-labels implies that the model is sufficiently capable of estimating the poses of said samples. The cross-domain setting makes it difficult to obtain high-quality labels. The challenge is compounded by additional gaps like different subjects anatomical sizings, different annotation standards across datasets, etc. Instead of estimating labels for the given target data, what if we try to reconstruct the data itself? This latter strategy is often used in self-supervised pose estimation, where learning is guided by reconstruction losses [33, 11, 34]. The challenge lies in linking the reconstruction scheme back to 37th Conference on Neural Information Processing Systems (Neur IPS 2023). the underlying pose1. While it is feasible to use a differentiable renderer, there are several drawbacks, such as the complexity of the renderer [23] and the difficulty of model fitting [16]. With this motivation, we design a simple yet effective reconstruction scheme that connects keypoint prediction via bone maps. Specifically, given a base and driving image pair with different poses, we estimate the keypoints and region-wise geometric transformations between the two poses, where each region corresponds to a bone linking two joints. The transformation is then applied with the predicted mask to obtain fine-grained warping (as shown in Fig. 1). The complete reconstruction is conducted in the generation module, which translates the base image into the driving image, specifically performing feature warping and inpainting in downsampling and upsampling structures. Intuitively, the more accurate the estimated keypoints and geometric transformation in the image pair, the better the reconstruction. Differences between the reconstruction and the original driving image can serve as a supervisory signal to improve keypoint localization. A key benefit of the reconstruction scheme is that it can leverage almost all the samples in the unlabelled target dataset and give refinement according to the reconstruction loss. Pseudo-labeled samples are often selected based on some confidence measure like heatmap response, but given that most neural networks are poorly calibrated [22, 19], the confidence is often unsatisfactorily aligned with pose accuracy. Many samples can be highly confident yet inaccurate and the predictions remain inaccurate in this way. This is a key reason why pseudo-labelling methods easily saturate (see Table 3 (b)). In contrast, our reconstruction scheme requires only the two images containing the same object (hand/human) and a rigid background. As such, we are able to improve those confident but wrong keypoints based on 2D evidence in the target domain that arises from the reconstruction loss (see Table 3 (c)). Also, for keypoints that are both confident and accurate, the reconstruction will not degrade the result since it has perfectly provided useful information for reconstruction. The supervision through reconstruction is indirect and is a weaker source of information than pose pseudo-labels. However, given that the two approaches are complementary, they can also be used in conjunction, i.e., a limited but strong supervision from pseudo-labelling in combination with a larger and diverse source of weak supervision from the reconstruction. To that end, we additionally propose a new selection criteria for pseudo-labels that integrates 2D evidence and 3D kinematic constraints. In combination, we are able to outperform state-of-the-art domain adaptation methods by 8%, with significant improvements on difficult keypoints, like fingertips for the hand, and head for humans by 29.9%. Our contributions can be summarized as A novel reconstruction-based approach for domain adaptive pose estimation. We introduce a reconstruction method to make up for the limitations of the pseudo-labelling, and realize the connection between image reconstruction and keypoint prediction by building bone maps, which enables the correction of keypoints through optimizing reconstruction loss. An improved pseudo-labelling selection criteria that integrates both 2D evidence and 3D kinematic constraints. The new criteria improves the quality of the pseudo-labels by enlarging the set of usable pseudo-labels by 13%. Domain adaptation results from synthetic to real-world data for both hand and human body show that we outperform state-of-the-art by a large margin, with as much as 29.9% improvement on difficult keypoints like fingertips and head and an overall 8% improvement. 2 Related Work 2.1 Cross-domain Pose Estimation Cross-domain challenges in pose estimation, especially in the context of synthetic-to-real, are best known for animal pose estimation [1, 20, 17], where it is challenging to collect and label sizable training data. Similar challenges, however, also exist for the human body [40] and hand [36, 18, 13], as deep learning algorithms become more data-hungry. The standard strategy is to apply consistency constraints [20, 36] or pseudo-labelling [17, 15] to learn effective features for the target domain. Consistency constraints can be applied in equivariant augmentations [20], and cross-modality [36]. 1We make a distinction between data synthesis and reconstruction. The former generates image observations based on given poses; the latter recreates image observations without knowledge of the underlying pose. To generate less-noisy pseudo-labels and guide the network fine-tuning, [15] utilized the normalized output from a teacher network and [1] proposed a progressive self-paced pseudo-label updating strategy to make the guidance more stable. Both works denoise pseudo-labels based on the heatmap activations while [36, 18] reduced noise with a pose correction procedure. Either strategy alone can lead to the neglect of additional correlation of depth or the accumulation of prediction errors. We therefore leverage both strategies simultaneously and embrace their union of pseudo-labels indicators when their consensus is high to enrich guidance on the target domain. 2.2 Conditional Image Reconstruction Reconstructing images conditioned on paired images is an effective technique to concentrate on specific attributes (e.g., keypoints) while ignoring irrelevant ones in pose estimation, where the focus can be summarized on conditions design [10, 11, 25]. Jakab et al. [10] proposed to disentangle keypoint heatmaps from appearance. The heatmaps, as conditions to generate images, can be optimized by the reconstruction loss. A follow-up work [11] utilized a skeleton image to facilitate the generation process. Furthermore, recent work [25] regressed affine parameters to deform a pre-defined shape template as conditions. However, they all extract appearance and keypoints from different images separately and overlook the geometric relationship across images, which can provide explicit guidance on image reconstruction. Using geometric relationships across images to perform reconstruction has been seen in image animation [27, 26, 28]. They predict motion-specific keypoints [27, 26] or semantic regions [28], and obtain transformation parameters to deform one image to resemble another image. Those motionspecific keypoints or region-based methods are generalizable in predicting transformation parameters and modeling complex movements. But they target achieving the mean and deviation of each region, which is unrelated to our purpose of joint locations. To emphasize the geometric relationship (i.e., the transformations between the two images) to perform reconstruction for refining joint locations, we connect semantic regions and joint locations via bone maps, which enables the correction of poses through optimizing the reconstruction loss. Naturally, the reconstruction becomes a complement to pseudo-labelling, and both of them contribute to domain adaptive pose estimation. 3.1 Problem Formulation Consider labelled data Ds = {(Isi, yi)}Ns i=1 from a source domain s and unlabelled data Dt = {(Iti)}Nt i=1 from target domain t, where I denotes an RGB image, y the pose label, and Ns and Nt are the number of samples in the source and target datasets. The objective is to learn a pose estimation system fkp = (fenc fdec), trained on {Ds, Dt} and apply it to hold-out test data on the target domain t for estimating 2D or 3D keypoints. We follow other domain adaptive pose estimation methods, and pre-train on the labelled source domain (synthetic) and then fine-tune on the unlabelled target domain (real-world). In pre-training, (Sec. 3.2), we use the labelled synthetic data Ds to learn an initial pose estimation network. For fine-tuning on unlabelled real-world data Dt (Sec. 3.3), we introduce a pseudo-labelling strategy to correct 2D pseudo-labels via 3D constraints in Sec. 3.3.1. Additionally, we propose a geometric reconstruction to correct keypoint locations from image signals in Sec.3.3.2. The overall pipeline for fine-tuning is shown in Fig. 1. 3.2 Pre-training on the Source Domain Keypoint Predictor The keypoint predictor fkp = (fenc fdec) is first pre-trained on the synthetic source images Is to estimate a heatmap ˆhp and a relative depth ˆd. As the synthetic images have ground truth labels, fkp can be learned in a supervised manner with ground-truths (hgt p, dgt): Ds ( ˆhp hgt p 2 + λd ˆd dgt 2), (1) 𝐁!/#$%: base/driving bone map 𝐂#$% !: inpainting map 𝐃#$% !: displacement field 𝐀#$% !: transformation parameters 𝒑: normalized coordinates 𝝁!/#$%: keypoint locations : multiplication 𝐌: mask : deformation 𝑓'() 𝑓#') 𝐈#$% Integrated pseudo-labels 𝐃#$% ! ℎF𝒑#$% (a) Fine-tuning on Target Domain (b) Warping Estimation Module Figure 1: Illustration of pipeline and modules of fine-tuning. (a) Fine-tuning on target domain. Given the base and driving images sharing the same background but different poses, we forward them into keypoint predictor fkp = (fenc fdec) to produce keypoint heatmaps ˆhp and relative depth predictions ˆd. Afterward, our integrated pseudo-labelling P (See Sec. 3.3.1) is introduced to generate pseudo-labels, while the warping estimation module W and the generation module G (See Sec. 3.3.2) are used to reconstruct driving images from base images. (b) Warping estimation module. Bone maps derived from keypoints are used to calculate coarse transformation parameters via SVD and Expand. With the transformation and base images, a hourglass is used to generate inpainting map Cdrv b and region-wise displacement field Ddrv b that will work on upsampled feature maps in the generator G. where λd is a hyper-parameter that trades off depth loss and heatmap loss. The ground-truth heatmap hgt p is constructed as a 2D Gaussian centered at the ground-truth coordinate (ugt, vgt) and a standard deviation of 2. The ground-truth dgt is the depth relative to the root keypoint. We follow the same practice in [9] to regress 2.5D and lift it to 3D pose. To elaborate, we employ soft-argmax operation on ˆhp and then perform Hadamard product with depth value maps to derive the relative depth ˆd. The expected outcome of this operation yields the 2D coordinates, denoted as (ˆu, ˆv). Invariant Feature Learning Models trained only on synthetic data Ds are unlikely to perform well on target data Dt (see Source only in Table 1). One reason is the domain gap between source and target data; typically, synthetic data used for pose estimation is not fully photorealistic, so the model does not generalize well to real-world data [4, 31]. To prevent fkp from overfitting to source-specific representations, which is common [38, 21], we augment the source data to resemble target data and align the original and augmented source features using contrastive learning [3]. Specifically, we train a Style Net from [24], which transfers the style of the target domain to the source, while preserving the source content, i.e., the pose. The keypoint predictor fkp is then trained with the supervised keypoint loss and an alignment on features with the following loss: i=1 log e( zi,z+ i /τ) PB k=1 1[k =i](e( zi,zk /τ) + e( zi,z+ k /τ)) , (2) with batch size B, temperature hyperparameter τ = 0.5, and denotes the cosine similarity. In the above equation, the features (z, z+) are derived from the source and augmented source image. [15] also applies a Style Net to augment source samples but only supervises the network output. Our work extends the supervision to the latent features, as it can further benefit task-specific feature learning by building a consistent embedding space. It is feasible to also add target data into invariant feature learning with the style of source data, but we empirically find no obvious improvement. Therefore, we only perform alignment between source and augmented source features considering the limited memory. 3.3 Fine-tuning on the Target Domain The fine-tuning on the target domain consists of integrated pseudo-labelling and geometric reconstruction. The integrated pseudo-labelling extracts confident pesudo-labels with guidance from both 2D heatmaps and 3D constraints. Since the label can still be noisy, we also have a geometric reconstruction strategy that generates a driving image based on the base image and pose transformation, which provides supervision simply from 2D image signals to improve those confident yet inaccurate keypoints. 3.3.1 Integrated Pseudo-labelling A commonly used strategy to adapt source models to a target domain is pseudo-labelling. In ensuring that pseudo-labels are valid, existing works have leveraged the heatmap activations in ˆhp [15], or kinematic feasibility of the pose ˆy [36]. Both strategies have drawbacks; the former ignores correlations that may exist in ˆd, while the latter uses a greedy strategy that may accumulate errors down kinematic chains. With these considerations in mind, we propose an integrated pseudo-labelling strategy that combines the use of the 2D heatmap and the kinematic feasibility of 3D predictions. Specifically, given a predicted heatmap ˆhp, we generate a 2D confidence mask M2D RB K as indicators to select keypoints with heatmap activations larger than the threshold γ2D in each batch during training. The 2D pseudo-label ehp is the gaussian heatmap ˆhp centered on the predicted 2D coordinate, as defined in [15]. For the 3D pseudo-labels, the prediction ˆy = (ˆu, ˆv, ˆd) is lifted into 3D coordinates and rectified as pose ey with reasonable feasibility based on the pose correction [36]. The 3D confidence mask M3D RB K indicates the qualified keypoints based on the comparison of L2 norm between the 2D projected coordinates before and after correction on ˆy to the threshold γ3D. For the first T epochs of fine-tuning, the pseudo-labels ehp and ey generated by 2D evidence and 3D correction, respectively, are both not confident, so we require them to train independently with their corresponding mask. Then we integrate the confidence mask M2D and M3D to complement each other and enrich the training set as M = M2D M3D, which considers more pseudo-labels for adaptation to both 2D and 3D. Overall, we optimize the following loss, ( 1 |M2D| P M2D ˆhp ehp 2 + λy |M3D| P M3D ˆy ey 2 epoch < T 1 |M| P M( ˆhp ehp 2 + λy ˆy ey 2) epoch T , (3) where |M| indicates the number of elements in M and λy = 0.1 is a coefficient that trades off pseudo-labelling between 2D heatmap loss and 3D loss. We find out that as fine-tuning proceeds, the overlap of confidence masks M2D and M3D will increase, at which point a single strategy can also provide valuable indicators for the other strategy. 3.3.2 Geometric Reconstruction Although the pseudo-labelling strategy is improved by utilizing both 2D and 3D information, the amount of pseudo-labels is limited and some confident labels are still incorrect. This is because these labels are generated from the model calibrated on the source domain, we therefore propose a geometric reconstruction that solely exploits specific features in the target domain. The performance of reconstructions is weaker compared to high-level pseudo-labeling. However, it can serve as complementary supervision by including all target data and correcting confident yet inaccurate keypoints by low-level evidence. The geometric reconstruction consists of Warping Estimation Module W and Generation Module G, as shown in Fig. 1(a). Given a pair of images from target domain t containing human or hand {Ib, Idrv} sampled in a video, The aim is to reconstruct driving image ˆIdrv R3 H W from base image Ib R3 H W , where the image size is H W. To do so, the generation module G warps base image Ib based on specifications Ddrv b R H 4 2 and Cdrv b R1 H 4 estimated by the warping estimation module W: ˆIdrv = G(Ib; Ddrv b, Cdrv b), where [Ddrv b, Cdrv b] = W(ˆhb p, ˆhdrv p ). (4) In Eq. 4, Ddrv b and Cdrv b denote a region-wise displacement field and an inpainting map respectively. The displacement field Ddrv b is a dense vector field that specifies each pixel location in Idrv with its corresponding location in Ib. And Cdrv b is an attention map ranging from 0 to 1, where the G should do more inpainting on pixels with lower values. We apply a Perceptual Loss to penalize the semantic differences between the reconstructed image ˆIdrv and its original image Idrv to refine the keypoints. The details of each component (in bold) as follows. Warping Estimation Module W Given the heatmap hp estimated from fkp, we decode it into coordinates by using a soft-argmax followed by an expectation; this combination is the standard approach followed in coordinate or integral regression methods [9, 29, 7, 6]. Specifically, we obtain the k-th 2D keypoint locations µk in all K keypoints while keeping the pretrained heatmaps: p Ω p hp, where hp = ehp P p Ωehp . (5) Ωis the area of the heatmaps, which are generated by fkp. Following the coarse-to-fine pipeline, to obtain a coarse motion of each body part, we need to obtain the pose transformation parameters. Previous works [25, 27] use the network to do direct regression but proved to have a poor generalization [28]. We thus follow [28] to use singular value decomposition (SVD) to compute transformation parameters of the bone map with respect to a reference heatmap, which is whitened and has zero mean and identity covariance. However, different from [28], where the bone maps are predicted by the network, we generate the bone map according to keypoint locations µk and the kinematic tree. Formally, we assume there are J chains in the kinematic tree and we denote the parent and child node of the kinematic chain cj as µjp and µjc respectively. The value of bone map Bj at pixel p of the j-th kinematic chain can be formulated as bj(p) = e dj(p)/σ2, (6) where σ is a hyperparameter that controls the thickness of the edge and dj denotes the distance between the pixel p and the edge determined by µjp and µjc. Accordingly, the transformation parameters2 A of I can be obtained by 1 2 j , µj], where Uj Sj Vj = X p Ω bj(p)(p µj)(p µj)T where µj is the mean of µjp and µjc. By Eq. 7, we can get the transformations Ab ref and Adrv ref. Finally, the geometric transformation of bone j from driving to base image is calculated by Adrv b = Adrv ref(Ab ref) 1. To generate a dense shift to deform Ib to approach Idrv, we next convert J transformation parameters to a fine-grained displacement field along with background transformation parameters. Specifically, with the input of Ib, Bb and Bdrv, an hourglass network outputs J +1 masks Mj indicating one background and J bones. M0 corresponds to the background. Mj(p) denotes the possibility of p belonging to the j-th region when j > 0. To ensure that the possibilities of every channel sum up to one, softmax is applied across the channels. Then, the displacement field Ddrv b(p) R2, is generated by Ddrv b(p) = j=0 Mj(p) Aj | {z } Expand Through region-wise displacement field Ddrv b, the network gains insight into the specific locations on the base image that require deformation to match the driving image. Nonetheless, certain regions may be absent in the base image while present in the driving image. To identify these areas in need of inpainting by the network, we add an additional layer after the same hourglass to generate the inpainting map Cdrv b. 2We only need to rotate, translate, and scale to realize the transformation between the designed bone maps, which is different from the previous works based on affine transformation. Generation Module G We build G with downsampling and upsampling, interconnected through skip connections. Similar to [28], the feature map from the skip connection is warped by displacement field Ddrv b and then activated with inpainting map Cdrv b. We add the processed feature map to the 1 Cdrv b activated feature map of the previous layer, and the sum is passed to the next layer. In this way, G can generate driving images through feature warping and limited inpainting from a similar image without the need for extensive rendering. Perceptual Loss Following [14, 35], we apply a multi-scale perceptual ℓ1 loss with a pre-trained VGG-19 network. Mathematically, the reconstruction loss Lrec is formulated as i |ϕi(ˆIdrv) ϕi(Idrv)|, (9) where ϕi(I) is the intermediate output of i-th layer of VGG-19. Training Procedure The overall training of our framework is done in two stages. First is the pretraining on the source domain with supervised learning in Eq. 1 and invariant feature learning in Eq. 2 with weighting hyperparameter λa: Lpre-train = Lsup + λa Lalign. (10) The second stage finetunes in the target domain, with pseudo-labelling from Eq. 3 and geometric reconstruction from Eq. 9, together with supervision from the source domain to avoid over-fitting: Lfine-tune = Lsup + λr Lrec + λp Lpseudo, (11) where λr and λp are balancing hyper-parameters. We have a warmstart in the fine-tuning where Lsup and Lpseudo are first used and then add Lrec together to update the overall pipeline. 4 Experiments 4.1 Datasets & Evaluation Metric Hand Pose Datasets. For the source domain, we consider the synthetic dataset RHD [43] with 44k rendered images for training. For the unlabeled target domain, we consider real-world datasets H3D [41] and MVHand [39]; these have 11k/2k and 42k/42k training/testing splits respectively. For H3D, we consider only the subset with one-handed gestures. H3D and MVHand are both multi-view datasets; we treat each viewpoint as a video source. Human Pose Datasets. SURREAL [31] is a synthetic human dataset with 6 million annotations including keypoint locations. Human3.6M [8] annotates 3.6 million real-world indoor human poses. We follow protocol 1, using S1 and S5-S8 for training and S9 and S11 for testing. 3DPW [32] is a challenging outdoor dataset with 24 videos for training and 24 videos for testing. We obtain 2D keypoints by projecting the corresponding 3D ground truth. Evaluation Metric. The convention of domain adaptive pose estimation works [20, 15] is to evaluate 2D keypoint detection with Percentage of Correct Keypoint (PCK). We report PCK@0.05, which tabulates the percentage of correct predictions within the range of 5% of the input image size; a higher PCK indicates better performance. For 3D keypoint estimation, we evaluate with mean end-point-error (EPE); a lower EPE indicates better performance. 4.2 Implementation Details Our framework is shown in Fig. 1. We use Res Net-101 as the feature extractor fenc and deconvolutional layers as the decoder fdec to produce a heatmap of size 2K 64 64 for 2D and depth. Each input image is cropped around the hand or the human and resized to 256 256. The initial learning rate is 1e-4, and we used Adam optimizer and degraded the learning rate with 0.1 at steps 60 and 75. Pre-training is done with a batch size of 64 for 40 epochs; fine-tuning is done with a batch size of 32 for 80 epochs. The hyperparameters λa, λr, λp are set as 0.1. We compare with state-of-the-art domain adaptation methods: Reg DA [12], CC-SSL [20], Uni Frame [15], Semi Hand [36], and Dual Net [18]. Where possible, results are reported directly from the corresponding papers; otherwise, results are based on officially released code. In pre-training, RHD [43] and SURREAL [31] are Method RHD H3D RHD MVHand MCP PIP DIP Fin All MCP PIP DIP Fin All Source only (Lsup) 67.4 64.2 63.3 54.8 61.8 32.1 70.5 64.6 39.8 52.7 Semi Hand [36] - - - - 67.2 - - - - 56.3 Dual Net [18] - - - - 74.9 - - - - 68.9 CC-SSL [20] 81.5 79.9 74.4 64.0 75.1 - - - - 60.2 Reg DA [12] 79.6 74.4 71.2 62.9 72.5 - - - - 60.1 Uni Frame[15] 86.7 84.6 78.9 68.1 79.6 45.0 77.7 71.0 57.6 62.9 Ours 89.4 88.3 81.9 73.9 84.1 76.4 76.9 68.7 61.0 70.8 Table 1: Hand 2D keypoint detection. Our method has the best average PCK@0.05 on both datasets and the most significant improvements on MCP and Fin. Method SURREAL Human3.6M SURREAL 3DPW Head Sld Elb Wrist Hip Knee Ankle All Head Sld Elb Wrist Hip Knee Ankle All Source Only 51.3 69.4 75.4 66.4 37.9 77.3 77.7 67.3 40.9 59.2 65.6 64.2 64.4 56.4 39.8 57.9 CC-SSL [20] - 44.3 68.5 55.2 22.2 62.3 57.8 51.7 41.0 58.5 76.3 67.4 77.6 61.7 47.8 62.5 Reg DA [12] - 73.3 86.4 72.8 54.8 82.0 84.4 75.6 41.8 65.8 73.3 68.1 72.8 65.0 49.3 64.8 Uni Frame[15] 63.5 78.1 89.6 81.1 52.6 85.3 87.1 79.0 54.6 68.9 77.5 74.8 75.5 67.8 60.3 70.5 Ours 87.5 89.8 90.1 78.7 69.5 85.4 84.8 83.7 66.6 86.6 77.8 75.8 83.6 71.9 65.3 76.7 Table 2: The comparison on PCK@0.05 for 2D human keypoint detection. We achieve the best average performance on both datasets with the most significant improvements on the head. used as labeled source dataset Ds in hand and human pose estimation, respectively. In fine-tuning, the training sets of H3D [41], MVHand [39], Human3.6M [8] and 3DPW [32] are individually used as unlabeled target dataset Dt for each pose adaptation experiment. And we evaluate on the corresponding test set. 4.3 Comparison with the State-of-the-Art Hand Pose. Table 1 shows that for 2D keypoint detection on the hand, our method outperforms the previous best method by 5.7% and 12.6% on the H3D and MVHand datasets respectively. For more in-depth analysis, we also report the individual joint results averaged across the fingers i.e., MCP, PIP, DIP, Fin3. Adapting RHD H3D gives a consistent improvement, with increases of 3.1% to 8.5%. On RHD MVHand, ours still achieves a significant improvement in PCK on MCP, increasing from 45.0 to 76.4. Table 5 (b) shows that our method is state-of-the-art on 3D hand pose estimation for both RHD H3D and RHD MVHand. Especially notable is that we surpass Dual Net [18], as we use only RGB images, while they incorporate additional modalities like background masks and depth images. Human Body Pose. Table 2 shows, for SURREAL Human3.6M and SURREAL 3DPW that our method outperforms state-of-the-art Uni Frame [15] by 5.9% for Human3.6M and 8.8% for 3DPW. The biggest improvements come in the head, shoulders, and hips. In particular, we observe the hip problem arising from the gap in annotation convention between SURREAL and Human3.6M, so all methods (including ours) have lower accuracy on the hips. However, we outperform others by 32.1% on Human3.6M, indicating our geometric reconstruction can help the model less affected by the annotation gap. 4.4 Analysis Effectiveness of Components. We made two core contributions: geometric reconstruction and integrated pseudo-labelling. We here provide the gain of each component in Table 3 (a). LUni Frame means we use the pseudo-labelling strategy in [15]. The comparison with LUni Frame reveals the effectiveness of integrated pseudo-labelling with a 1.2 PCK improvement. Together with geometric reconstruction, we can further boost the increase to 2.7. Lheatmap means that we remove the warping estimation module and purely use heatmaps of both images to reconstruct ˆIdrv. The results show that using explicit guidance in the generation module could help effectively link reconstruction back to the underlying poses. 3Hand joints moving away from the palm: Meta Carpo Phalangeal (MCP), PIP (Proximal Inter Phalangeal), DIP (Distal Inter Phalangeal) and Fin (Fingertips). before after GT before after GT Figure 2: Comparisons before and after applying geometric reconstruction. Our method can correct the predictions that are on the background (left) and alleviate the annotation gap (right). Method RHD H3D SURREAL Human3.6M Lpre-train 67.2 69.0 +LUni Frame 80.2 79.8 +Lpseudo 81.4 80.8 +Lheatmap 81.6 81.2 +Lrec 84.1 83.7 80 82 84 86 88 90 Epoch Ours Uni Frame head shoulder hip others Type of Keypoints Accuracy of Pseudo-labels (PCK) (a) (b) (c) Table 3: (a) Ablation study on the proposed components and the reconstruction module contributes the most to the final performance. (b) Percentage of correct pseudo-labels in the training set of MVHand dataset. Both strategies saturate but the proposed integrated pseudo-labelling induces more correct labels. (c) Accuracy comparison before and after applying reconstruction loss Lrec for a specific set of confident keypoints on Human3.6M training set. After applying Lrec, we are able to improve the pseudo-label of those keypoints previously proposed using only Lpseudo and Lsup. Keypoint Correction by Geometric Reconstruction. We analyze the impact of geometric reconstruction on confident yet inaccurate keypoints. We first train the model with Lsup and Lpseudo until convergence, and then save the indices of confident keypoints. Next, the reconstruction loss Lrec is added to fine-tune the model. We show the accuracy of the pseudo-labels for the keypoints associated with the saved indices, both before and after adding the reconstruction loss in Table 3 (c). The reconstruction loss obviously improves the confident yet inaccurate keypoints (i.e., head, shoulder and hip), while having a negligible influence on other keypoints. This is because confident yet inaccurate keypoints will cause large reconstruction errors based on image evidence while others will not. Also, we observe that those confident yet inaccurate keypoints tend to be specific keypoints, and speculate it is due to the annotation gap. The Amount of Correct Pseudo-labels. We select pseudo-labels with high confidence. However, we cannot assure that their accuracy is within PCK@0.05 for effective adaptation. We therefore tabulate the percentage of correct pseudo-labels, whose accuracy is within PCK@0.05, in unlabeled training data aggregated by the last ten epochs. As shown in Table 3 (b), the percentage saturates to 59% for the pseudo-labelling strategy and to 72% for the proposed integrated pseudo-labelling. Since the pseudo-labels are explored on training data with augmentations, their accuracy might not align consistently with the accuracy observed on the evaluation set. It is evident that both our approach and Uniframe eventually reach a saturation point for accurate pseudo-labels, but our method involves more usable pseudo-labels and leads to a better result. Fairness & Stride Strategy. In our method, we need a pair of images with a similar background and the same object (human/hand). In contrast, existing state-of-the-art methods only require a single image during training on the target domain. For a fair comparison, we first consider the most common module to utilize adjacent paired input, i.e., temporal consistency by applying ℓ2 loss on the predicted heatmaps hi p hi+1 p 2 of two consecutive frames Ii, Ii+1. Results in Table 5 (a) show it can only gain marginal improvement. Another consideration is whether the difference between the two poses affects the effectiveness of the reconstruction. We attempted to amplify the pose difference of two images from small to large by adjusting the stride from [0,5] to [30, 50]. As shown in Table 4, our method performs best on [5,15]. This makes sense that degenerately close Stride [0, 5] [0, 15] [5, 15] [15, 30] [30, 50] SURREAL Human3.6M 81.8 83.5 83.7 83.1 82.5 Table 4: Ablation study on frame strides for the task of SURREAL Human3.6M evaluated by PCK@0.05. We can see that all strides lead to improvement, and our method is robust to the selection of strides within a certain range. Method RHD MVHand SURREAL Human3.6M Reg DA[12] 60.7( 0.6) 76.4( 0.8) Uni Frame[15] 63.4( 0.5) 79.3( 0.3) Ours 70.8 83.7 Method RHD H3D RHD MVHand Source only 27.77 21.21 Semi Hand[15] 19.19 19.75 Dual Net[18] 17.08 16.45 Ours 16.92 16.37 Table 5: (a) We add temporal consistency to ensure equivalence in the input information, which only brings minor improvement. (b) The comparison on EPE (mm) for 3D hand pose. By only improving 2D and applying a simple 3D pseudo-labelling strategy, we beat the state-of-the-art methods. frames will make the network find the shortcut to self-reconstruction and far frames will make the generation too difficult to have a meaningful perceptual loss on keypoint locations. 4.5 Qualitative Results Hand Pose. Fig. 2 (left) shows a common case in only using pseudo-labelling, where the fingertip is predicted on the background, although it is near the ground truth location. Applying our geometric reconstruction results in more accurate fingertip predictions. Besides, the wrong prediction of the pinky fingertip falling on the palm can also be rectified via the perceptual loss in our reconstruction module. Human Body Pose. Fig. 2 (right) shows the comparison of our method on a Human3.6M sample before and after applying geometric reconstruction, where the shoulders are much lower in before compared to GT . One possible cause is the difference in the annotation conventions across datasets, e.g. The annotation of the shoulders in SURREAL is closer to the chest, while Human3.6M localizes them at a reasonable shoulder position (See Supp.). Similar differences due to the annotation gap are reflected in the accuracy of head and hips, as shown in Table 2. Our method, however, can refine the pose prediction by leveraging the correlation between bone map and reconstruction, which is independent of the pose prior learned in the source domain. More analysis can be seen in Supp. 5 Conclusion Pseudo-labelling is a common strategy for domain adaptive pose estimation which selects confident predictions to supervise the network. However, the number of accurate labels under this selection will saturate with the progress of training and some of the wrong predictions will remain unchanged. We hereby introduce a novel geometric reconstruction module to relieve this problem that aims to reconstruct the driving image according to keypoint locations and geometric transformations from the base image to the driving image. The incredible improvement especially comes from keypoints that generalize extremely poorly to target domain by only pseudo-labelling. We hope this will inspire future work to explore reconstruction-based methods to bridge the gap between source and target domains. Acknowledgments This research / project is supported by the Ministry of Education, Singapore, under its MOE Academic Research Fund Tier 2 (STEM RIE2025 MOE-T2EP20220-0015). [1] Jinkun Cao, Hongyang Tang, Hao-Shu Fang, Xiaoyong Shen, Cewu Lu, and Yu-Wing Tai. Cross-domain adaptation for animal pose estimation. In ICCV, 2019. [2] Ching-Hang Chen and Deva Ramanan. 3d human pose estimation= 2d pose estimation+ matching. In CVPR, 2017. [3] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In ICML, 2020. [4] Wuyang Chen, Zhiding Yu, Zhangyang Wang, and Animashree Anandkumar. Automated synthetic-toreal generalization. In ICML, 2020. [5] Salehe Erfanian Ebadi, Saurav Dhakad, Sanjay Vishwakarma, Chunpu Wang, You-Cyuan Jhang, Maciek Chociej, Adam Crespi, Alex Thaman, and Sujoy Ganguly. Psp-hdri +: A synthetic dataset generator for pre-training of human-centric computer vision models. In ICML, 2022. [6] Kerui Gu, Linlin Yang, Michael Bi Mi, and Angela Yao. Bias-compensated integral regression for human pose estimation. TPAMI, 2023. [7] Kerui Gu, Linlin Yang, and Angela Yao. Removing the bias of integral pose regression. In ICCV, pages 11067 11076, 2021. [8] Catalin Ionescu, Dragos Papava, Vlad Olaru, and Cristian Sminchisescu. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. TPAMI, 36(7):1325 1339, 2013. [9] Umar Iqbal, Pavlo Molchanov, Thomas Breuel Juergen Gall, and Jan Kautz. Hand pose estimation via latent 2.5d heatmap regression. In ECCV, 2018. [10] Tomas Jakab, Ankush Gupta, Hakan Bilen, and Andrea Vedaldi. Unsupervised learning of object landmarks through conditional image generation. In Neur IPS, volume 31, 2018. [11] Tomas Jakab, Ankush Gupta, Hakan Bilen, and Andrea Vedaldi. Self-supervised learning of interpretable keypoints from unlabelled videos. In CVPR, pages 8787 8797, 2020. [12] Junguang Jiang, Yifei Ji, Ximei Wang, Yufeng Liu, Jianmin Wang, and Mingsheng Long. Regressive domain adaptation for unsupervised keypoint detection. In CVPR, 2021. [13] Rui Jin, Jing Zhang, Jianyu Yang, and Dacheng Tao. Multibranch adversarial regression for domain adaptative hand pose estimation. TCSVT, 32(9):6125 6136, 2022. [14] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and superresolution. In ECCV. Springer, 2016. [15] Donghyun Kim, Kaihong Wang, Kate Saenko, Margrit Betke, and Stan Sclaroff. A unified framework for domain adaptive pose estimation. In ECCV. Springer, 2022. [16] Nikos Kolotouros, Georgios Pavlakos, Michael J Black, and Kostas Daniilidis. Learning to reconstruct 3d human pose and shape via model-fitting in the loop. In ICCV, pages 2252 2261, 2019. [17] Chen Li and Gim Hee Lee. From synthetic to real: Unsupervised domain adaptation for animal pose estimation. In CVPR, 2021. [18] Qiuxia Lin, Linlin Yang, and Angela Yao. Cross-domain 3d hand pose estimation with dual modalities. In CVPR, 2023. [19] Charlotte Loh, Rumen Dangovski, Shivchander Sudalairaj, Seungwook Han, Ligong Han, Leonid Karlinsky, Marin Soljacic, and Akash Srivastava. On the importance of calibration in semi-supervised learning. In ar Xiv preprint ar Xiv:2210.04783, 2022. [20] Jiteng Mu, Weichao Qiu, Gregory D Hager, and Alan L Yuille. Learning from synthetic animals. In CVPR, 2020. [21] Hyeonseob Nam, Hyun Jae Lee, Jongchan Park, Wonjun Yoon, and Donggeun Yoo. Reducing domain gap by reducing style bias. In CVPR, pages 8690 8699, 2021. [22] Paweł A Pierzchlewicz, R James Cotton, Mohammad Bashiri, and Fabian H Sinz. Multi-hypothesis 3d human pose estimation metrics favor miscalibrated distributions. In ar Xiv preprint ar Xiv:2210.11179, 2022. [23] Neng Qian, Jiayi Wang, Franziska Mueller, Florian Bernard, Vladislav Golyanik, and Christian Theobalt. Html: A parametric hand texture model for 3d hand reconstruction and personalization. In ECCV, 2020. [24] Dan Ruta, Saeid Motiian, Baldo Faieta, Zhe Lin, Hailin Jin, Alex Filipkowski, Andrew Gilbert, and John Collomosse. Aladin: all layer adaptive instance normalization for fine-grained style similarity. In ICCV, 2021. [25] Luca Schmidtke, Athanasios Vlontzos, Simon Ellershaw, Anna Lukens, Tomoki Arichi, and Bernhard Kainz. Unsupervised human pose estimation through transforming shape templates. In CVPR, pages 2484 2494, 2021. [26] Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. Animating arbitrary objects via deep motion transfer. In CVPR, 2019. [27] Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. First order motion model for image animation. In Neur IPS, 2019. [28] Aliaksandr Siarohin, Oliver J Woodford, Jian Ren, Menglei Chai, and Sergey Tulyakov. Motion representations for articulated animation. In CVPR, pages 13653 13662, 2021. [29] Xiao Sun, Bin Xiao, Fangyin Wei, Shuang Liang, and Yichen Wei. Integral human pose regression. In ECCV, 2018. [30] Alexander Toshev and Christian Szegedy. Deeppose: Human pose estimation via deep neural networks. In CVPR, 2014. [31] Gul Varol, Javier Romero, Xavier Martin, Naureen Mahmood, Michael J Black, Ivan Laptev, and Cordelia Schmid. Learning from synthetic humans. In CVPR, 2017. [32] Timo Von Marcard, Roberto Henschel, Michael J Black, Bodo Rosenhahn, and Gerard Pons-Moll. Recovering accurate 3d human pose in the wild using imus and a moving camera. In ECCV, 2018. [33] Chengde Wan, Thomas Probst, Luc Van Gool, and Angela Yao. Self-supervised 3d hand pose estimation through training by fitting. In CVPR, 2019. [34] Chengde Wan, Thomas Probst, Luc Van Gool, and Angela Yao. Dual grid net: Hand mesh vertex regression from single depth maps. In ECCV, pages 442 459. Springer, 2020. [35] Wei Wang, Xavier Alameda-Pineda, Dan Xu, Pascal Fua, Elisa Ricci, and Nicu Sebe. Every smile is unique: Landmark-guided diverse smile generation. In CVPR, pages 7083 7092, 2018. [36] Linlin Yang, Shicheng Chen, and Angela Yao. Semihand: Semi-supervised hand pose estimation with consistency. In ICCV, 2021. [37] Linlin Yang and Angela Yao. Disentangling latent hands for image synthesis and pose estimation. In CVPR, 2019. [38] Yalan Ye, Ziwei Huang, Tongjie Pan, Jingjing Li, and Heng Tao Shen. Reducing bias to source samples for unsupervised domain adaptation. NN, 141:61 71, 2021. [39] Ziwei Yu, Linlin Yang, Shicheng Chen, and Angela Yao. Local and global point cloud reconstruction for 3d hand pose estimation. In BMVC, 2021. [40] Xiheng Zhang, Yongkang Wong, Mohan S Kankanhalli, and Weidong Geng. Unsupervised domain adaptation for 3d human pose estimation. In ACMMM, pages 926 934, 2019. [41] Zhengyi Zhao, Tianyao Wang, Siyu Xia, and Yangang Wang. Hand-3d-studio: A new multi-view system for 3d hand reconstruction. In ICASSP, 2020. [42] Christian Zimmermann, Max Argus, and Thomas Brox. Contrastive representation learning for hand shape estimation. In GCPR, 2021. [43] Christian Zimmermann and Thomas Brox. Learning to estimate 3d hand pose from single rgb images. In ICCV, 2017.