# dive_deeper_into_integral_pose_regression__1d15d0b9.pdf Published as a conference paper at ICLR 2022 DIVE DEEPER INTO INTEGRAL POSE REGRESSION Kerui Gu1, Linlin Yang1,2 & Angela Yao1 1National University of Singapore, Singapore 2University of Bonn, Germany {keruigu, yangll, ayao}@comp.nus.edu.sg Integral pose regression combines an implicit heatmap with end-to-end training for human body and hand pose estimation. Unlike detection-based heatmap methods, which decode final joint positions from the heatmap with a non-differentiable argmax operation, integral regression methods apply a differentiable expectation operation. This paper offers a deep dive into the inference and back-propagation of integral pose regression to better understand the differences in performance and training compared to detection-based methods. For inference, we give theoretical support as to why expectation should always be better than the argmax operation, i.e. integral regression should always outperform detection. Yet, in practice, this is observed only in hard cases because the heatmap activation for regression shrinks in easy cases. We then experimentally show that activation shrinkage is one of the leading causes for integral regression s inferior performance. For backpropagation, we theoretically and empirically analyze the gradients to explain the slow training speed of integral regression. Based on these findings, we incorporate the supervision of a spatial prior to speed up training and improve performance. 1 INTRODUCTION 2D human pose estimation aims to detect the image coordinates of the body and/or the hand. In recent years, detection-based methods (Newell et al., 2016; Xiao et al., 2018; Sun et al., 2019; Li et al., 2019) and integral pose regression methods (Sun et al., 2018; Iqbal et al., 2018) have emerged as two common paradigms for human pose estimation. Both methods learn a heatmap representing pixel-wise likelihoods of the joint positions. The heatmap is learned explicitly for detection-based methods but remains implicit or latent for integral regression methods. To decode the heatmaps to joint coordinates, detection methods use an argmax, while integral regression methods take the expected value. As the expectation operation is differentiable, integral regression has the benefit of being end-to-end learnable, even though detection methods seem more competitive accuracywise (COC, 2020). Gu et al. (2021) s recent work showed a curious performance difference between integral regression and detection methods. With a nuanced split over the evaluation set, integral regression outperforms detection when test samples are harder , i.e. with fewer keypoints present in the scene, under higher occlusion and lower bounding box resolutions. Given that both detection and integral regression methods work with fully convolutional feed-forward architectures, the questions naturally arise: What are the reasons behind this performance difference? Why is integral regression able to excel in these hard cases, when its overall performance seems to lag behind detection methods (COC, 2020)? These questions serve as the motivation for our closer study and analysis of integral pose regression. Detection and integral regression methods differ in two aspects in both the forward and backward pass (see Fig. 1 for an overview). In the forward pass, the heatmap is decoded with an argmax for detection whereas a softmax normalization and expectation are used for integral regression. In the backward pass, detection methods are supervised with an explicitly defined Gaussian heatmap centered on the ground truth joint, while integral regression is supervised by the joint coordinates directly. Through detailed theoretical analysis and experimentation on the decoding and backpropagation process, we make the following findings and contributions: Published as a conference paper at ICLR 2022 1. We propose a unified model of the heatmap to interpret and compare detection and integral regression methods. We verify the model experimentally and show that as samples shift from hard to easy, the activation region on the heatmap shrinks for both detection and integral regression methods. 2. We demonstrate experimentally that degenerately small regions of activation degrade the accuracy of both detection and integral regression methods. 3. Integral regression methods, which decode the heatmap with an expectation operation, should result in a lower expected end-point error than detection methods that decode with an argmax operation. In practice, this can be observed only for hard samples due to the shrinkage of the active region on the heatmap. 4. Direct supervision with the joint coordinates in integral regression, although end-to-end, suffers from gradient vanishing and provides less spatial cues for learning the heatmap than the explicit heatmap supervision of detection methods. As a result, the training of integral regression is more inefficient and slower to converge than detection methods. Our findings provide insight into integral pose regression, which has better theoretical performance, and show that the density of heatmaps plays an important role in decoding heatmaps to coordinates. 2 RELATED WORK Since the concept of heatmaps were proposed in (Tompson et al., 2014), detection-based methods have been top performers in human pose estimation. Existing detection-based works emphasize extracting high-quality multi-resolution features. Particularly, Xiao et al. (2018) proposed to adapt Res Net with deconvolution layers while Hourglass (Newell et al., 2016) and Cascaded Pyramid Network (CPN) (Li et al., 2019) introduced cascaded network architectures with a coarse-to-fine design paradigm. High-Resolution Network (HRNet) (Sun et al., 2019) follows the coarse-to-fine paradigm and further improves performance by adopting more dense connections across different resolution representations. The integral regression method of Sun et al. (2018) introduced a competitive regression-based framework but is surpassed by more advanced detection-based works (Xiao et al., 2018; Sun et al., 2019). Numerical regression-based methods, which directly regress joint coordinates, are commonly used in facial landmark detection (Feng et al., 2018; Zhu et al., 2020). These methods, however, are not as accurate as detection-based methods on human pose estimation. In order to improve the accuracy, integral regression methods attempt to implicitly merge the spatial knowledge from the heatmap via a latent heatmap. Integral regression methods are especially preferred in hand pose estimation (Spurr et al., 2020; Yang et al., 2021) but are still less common for human pose estimation (Sun et al., 2018; Nibali et al., 2018). Recently, two parallel lines of work have emerged to analyse detection and integral regression methods separately. For detection-based methods, Huang et al. (2020) identified a heatmap bias caused by the inconsistency of the coordinate system transformations and solved it by redesigning the transformations during data processing. Additionally, as the predicted heatmaps during inference may violate the Gaussian assumption and worsen the performance, Huang et al. (2020) proposed to find the optimal Gaussian location based on the prediction while Zhang et al. (2020) modulated the predicted Gaussian distribution for decoding. For integral regression methods, Nibali et al. (2018) experimentally compared different heatmap regularizers, heatmap normalization schemes and loss functions. They found that using Jensen Shannon regularization on the heatmap with a softmax normalization and L1 loss achieves the best performance for integral regression. More recently, Gu et al. (2021) performed a systematic comparison of detection and integral regression methods with a common backbone and discovered the performance advantage of integral regression on the hard samples. Furthermore, Gu et al. (2021) revealed a bias that arises from taking the expectation after the softmax normalization. As such, they proposed a compensation scheme to mitigate the bias, which in turn improves the overall performance of integral regression methods, making them competitive with detection methods. Published as a conference paper at ICLR 2022 Figure 1: Comparison of the two decoding processes for detection-based methods (pink panel) and integral regression methods (green panel). They share the same encoding (yellow panel) but have different representations of the ground truth (gray panel). 3 PRELIMINARIES ON HUMAN POSE ESTIMATION In this work, we target the more commonly used top-down form of pose estimation in which a person detector already provides a cropped image I of the person. For simplicity, we focus our discussion on one given joint out of the K total joints in the body. The pose estimation model outputs a heatmap ˆH RM N, where M and N are the dimensions of the spatial heatmap. Typically, M and N are scaled down by a factor of 4 from the original input dimensions of I (Xiao et al., 2018; Sun et al., 2019). The heatmap ˆH represents a (discrete) spatial likelihood P(J|I), where J R1 2 is the 2D coordinates of the joint. In practice, all K heatmaps are predicted simultaneously by the same network, where each joint is one channel. In both detection and integral regression methods, the coordinates ˆJ are decoded from ˆH, where the two methods differ in the manner of decoding (see Sec. 3.1) and the form of supervision applied (see Sec. 3.2). 3.1 HEATMAP DECODING: MAX VERSUS EXPECTED VALUE Detection methods apply an argmax on ˆH indexed by p to estimate the joint coordinates ˆJde: ˆJde = arg max p ˆH(p). (1) Taking an argmax can be interpreted as taking a maximum likelihood on the heatmap ˆH, assuming that ˆH is proportional to the likelihood. In practice, the final ˆJde value is determined as a linear combination of the highest and second-highest response on ˆH as a way to account for quantization effects in the discrete heatmap (Newell et al., 2016). A more recent work, DARK (Zhang et al., 2020), approximates the true prediction by a Taylor series evaluated at the maximum activation of the heatmap and shows this to be more accurate. Integral pose regression applies an expectation operation on ˆH to give a mean estimate of the joint coordinates. To do so, the heatmap must first be normalized to sum up to 1; the most common and effective approach (Nibali et al., 2018) is to apply a softmax normalization. Afterwards, the predicted joint ˆJre with x and y components ˆJx and ˆJy1 is determined by taking the expectation on the normalized heatmap H with elements hp at location p: ˆJre = ˆJx ˆJy p Ω p hp where hp = eβˆhp P (p ) Ω eβˆhp . (2) Here, Ωis the domain of the heatmap and β is a scaling parameter used in the softmax normalization. Note that softmax normalization assigns a non-zero value to all pixels in H, even if it was originally zero in ˆH. These values also contribute to the expected value, resulting in a center-biased estimated 1For clarity, we drop the subscript re , as we refer to the individual components for integral regression only. Published as a conference paper at ICLR 2022 joint coordinate ˆJre (Gu et al., 2021). The smaller the β, the stronger the bias. Although Gu et al. (2021) proposed a compensation scheme, for the purpose of our analysis, we will assume that β is sufficiently large such that the impact of the bias is negligible. 3.2 SUPERVISION: EXPLICIT HEATMAP VERSUS GROUND TRUTH COORDINATES Detection methods are learned by providing supervision on the heatmap. The ground truth H is given as a circular Gaussian, with a mean centered at the ground truth joint coordinate (see Fig. 1). The loss applied is a pixel-wise MSE between the ground truth hp and the predicted ˆhp: Lde = ||H ˆH||2 2 = X p Ω (hp ˆhp)2, (3) where Ωis the domain of the heatmap. As the loss is defined in terms of the heatmap and not in terms of the predicted joint coordinates, detection-based methods are not end-to-end in their learning, and this is often cited as a drawback (Sun et al., 2018; Zhang et al., 2020). Integral pose regression defines a loss based on the difference between the prediction ˆJre and the ground truth joint location J. The L1 loss empirically performs better than L2 (Sun et al., 2018): Lre = Jgt ˆJre 1 = (| ˆJx Jx| + | ˆJy Jy|). (4) Integral regression methods are end-to-end because they provide supervision at the joint level. As ˆH is learned only implicitly, some works refer to the heatmap as latent (Iqbal et al., 2018). 3.3 PERFORMANCE DIFFERENCES Sun et al. (2018) and Gu et al. (2021) compared the performance of detection and integral regression methods on human pose estimation. Sun et al. (2018) found that integral regression is either as competitive or better for 2D pose estimation, though results are not fully conclusive, as the backbone they used for integral regression is different from detection. The more systematic and detailed comparison from Gu et al. (2021) used a common backbone. It also partitioned the data based on three factors of variation dominating current human pose estimation benchmarks (Ruggero Ronchi & Perona, 2017): the number of joints present in the scene, the percentage of occlusion and the largest dimension of the bounding box input. The hard samples are those with either 1-5 joints present, > 50% occlusion, or 32-64px input. For these samples, integral regression had, on average, 15% lower end-point error (EPE) than detection. The easy samples are defined by 11-17 joints, <10% occlusion, or >128px input. In these cases, detection methods were marginally better. The remaining cases with factors in between were considered to be medium ; performance of both methods was roughly equivalent on these samples. 4 ANALYSIS ON HEATMAP DECODING 4.1 LOCALIZED HEATMAP MODEL To analyze the performance differences between detection and integral regression methods, we make the following assumptions on the heatmap itself. First and foremost, we assume that a well-trained network will produce a heatmap ˆhp with significant or large values in a localized support region around ˆJ. We denote the region of support as Φ and assume that outside of Φ, the heatmap activation is approximately zero (see Fig. 2 (a)). Specifically, the normalized heatmap2 H is modelled by some density distribution P. We make no assumption on the form of P, but do assume that it fully captures the support for the joint location within Φ and that Φ is centered on the expected value of P on location µ = (µx, µy) with an area of (2s + 1, 2s + 1), i.e. p Ω p hp = X p Φ(µ,s) p hp = j=µy s i hij, µx+s P j=µy s j hij 2For the purpose of discussion, we may consider normalizing the heatmap for detection methods as per Eq. 2, which will not affect the outcome since arg max( H)=arg max( ˆH). Published as a conference paper at ICLR 2022 (a) Illustration of Φ (b) Activation sum of H (c) Optimal σ of predicted heatmaps (d) AP versus ground truth σ Figure 2: (a) The heatmap is only activated within Φ; (b) more heatmap activations are bounded by a (2s + 1) (2s + 1) region for integral regression than detection; (c) if the heatmap is approximated as a Gaussian, the resulting σ shrinks as samples shift from hard to easy; this shrinkage is faster and more extreme for integral regression compared to detection; (d) in detect methods, the optimal σ is around 2; smaller and larger σ hurt performance. Experimental Verification of Localized Heatmap Model The assumption of localized activations is reasonable for detection methods as their heatmaps are learned explicitly to match a (localized) ground truth Gaussian. We empirically verify this model by tallying the activations in the normalized heatmap for both detection and integral regression methods based on the easy/medium/hard splits as specified in Sec. 3.3. Specifically, we train both a detection and an integral regression network for human pose estimation. For both networks, we use a Simple Baseline (SBL) (Xiao et al., 2018) architecture with a Res Net50 (He et al., 2016) backbone. The two networks differ only in their manner of decoding and form of supervision as outlined in Section 3. We then apply the networks on the standard human pose estimation benchmark MSCOCO (Lin et al., 2014) using the same easy, medium and hard splits as Gu et al. (2021). For each produced heatmap of 64 48, we estimate an activation sum A by summing the normalized heatmap activations within a (2s+1) (2s+1) square around the ground truth coordinate, (Jx, Jy): j=Jy s hij. (6) From Fig. 2 (b), we show that with a sufficiently large s, e.g. 9, the activation sum exceeds 95%. 4.2 EXTREME LOCALIZATION AS A CAUSE FOR PERFORMANCE DIFFERENCES Expected End-Point Error (EPE) For detection methods, which decode the heatmap with an argmax, we assume that each position p has some probability w(µ, s) of being the maximum activation, i.e. arg max(P)Φ(µ,s) w(µ, s). Here, w(µ, s) represents a radially symmetric distribution centered on µ with all non-zero support contained within Φ. It follows that the expected EPE for one sample of a detection method can be defined as: Ede(µ, s) = X p Φ(µ,s) w(p)||Jgt p||2, (7) where Jgt denotes the ground truth coordinates. For integral regression methods, which decode the heatmap with an expectation, the estimated joint coordinate aligns with the center of Φ by definition, i.e. ˆJ=µ, leading to the following expected EPE for a single sample: Ere(µ) = ||Jgt µ||2. (8) It can be verified that the Ede Ere (see proof in Appendix A.2). This result suggests that for some fixed (µ, s), the expected EPE of regression should always be better than detection. However, the results of Gu et al. (2021) as summarized in Sec. 3.3 have clearly shown otherwise. This inconsistency can be better understood with the following two experiments. Experimental Verification: Integral Regression Methods Are More Localized than Detection The plot in Fig. 2 (b) indicates that regression methods seem to be more localized than detection methods. In particular, for smaller s, A is smaller for detection methods than integral regression methods. Published as a conference paper at ICLR 2022 Model Detection Regression +σ=0.5 +σ=1 +σ=2 +σ=3 +σ=4 AP 71.3 67.1 65.1 68.5 71.4 69.8 66.9 APe/m 74.3 70.4 68.2 71.9 74.1 73.0 70.2 APh 43.2 44.7 43.9 45.2 47.4 47.6 46.9 A(s = 2)e/m 0.49 0.65 0.75 0.63 0.54 0.49 0.45 A(s = 2)h 0.30 0.35 0.37 0.36 0.41 0.39 0.39 Table 1: Performance and activation sum of detection baseline, regression baseline, and regression baseline added KLDiv loss with different σs. When σ = 0.5, 1, the results are not improved but even worsened. When σ=2, the spread of heatmaps is smoothed and the results even exceed detection baselines. To further verify, we compare the heatmaps with an idealized Gaussian heatmap of varying standard deviations or σ. We plot the σ that gives the lowest Pearson Chi-square statistic, i.e. the highest similarity for the different levels of difficulty in Fig. 2 (c) (detailed values are provided in Appendix A.3). As expected, the optimal σ decreases as the samples progress from hard to easy; this result is in line with Fig. 2 (b) and shows that heatmaps are more localized for easy samples. However, we also observe that the optimal σ for intgral regression is much smaller for the medium and easy cases. We speculate that this difference arises because detection methods are trained with a Gaussian heatmap of σ = 2 (Xiao et al., 2018; Sun et al., 2019). Integral regression methods, on the other hand, have no such explicit supervision, and therefore no restriction on the extent of localization. Experimental Verification: Extremely Localized Heatmaps Degrade Performance Is it possible that smaller hypothetical σ, i.e. more localized heatmaps, causes poor performance? We verify this for detection methods by applying ground truth Gaussian heatmaps of varying σ for training the networks as per Sec. 4.1 and evaluate network performance using Average Precision (AP). A higher AP corresponds to more accurately located joints. Fig. 2 (d) confirms that σ = 2 is optimal. A smaller σ = 1 degrades the performance slightly, and there is a significant drop in the extreme case when σ=0.5. To verify the impact of σ on regression methods, we add a distribution prior to the heatmap to prevent the shrink of the support region Φ. Specifically, we apply a Kullback-Leibler Divergence (KLDiv) loss between the non-normalized heatmap ˆH and a Gaussian heatmap with varying σ. From Table 1, we see that the regression baseline is worse than detection baseline. However, when we add a prior on the heatmap distribution to encourage a sufficiently large Φ, i.e. (+σ = 1, 2, 3), performance improves. At the optimal +σ = 2, integral regression becomes competitive with and even exceeds the detection baseline with the optimal σ. Most importantly, extreme σ values cause performance to degenerate, especially in the case of the extremely small +σ=0.5. We also demonstrate the performance and localization extent of easy/medium and hard cases by AP and activation sum A(s = 2). After adding priors (e.g. , +σ = 2), A(s = 2)e/m decreases from 0.65 to 0.54, indicating the heatmaps are less localized, and the performance APe/m is significantly improved. We provide the detailed training information regarding different loss weights for better understanding in Appendix A.4 and detailed changes in heatmap shrinkage in Appendix A.5. 5 ANALYSIS OF SUPERVISION AND LEARNING Given the same heatmap, different methods of heatmap decoding not only affect the final coordinates but also determine the gradients that arise from the loss. The gradients in turn influence the learning and thereby change the generation of the heatmaps. In this section, we explore the gradients of heatmaps with respect to the loss functions for detection and integral regression methods and pinpoint the specific gradient components that slow down the learning of integral regression. Published as a conference paper at ICLR 2022 5.1 HEATMAP GRADIENTS Detection Methods The gradient of Lde (see Eq. 3) with respect to each pixel in the estimated heatmap ˆhp is straightforward: Lde ˆhp = 2(ˆhp hp). (9) The gradient in Eq. 9 features the predicted heatmap value ˆhp and the ground truth heatmap value hp. It provides explicit supervision at every pixel. False positives and false negatives are penalized by reducing wrong high likelihoods and raising incorrect low likelihoods, respectively. For each position p in the heatmap, if the predicted ˆhp is smaller than the ground truth hp, the gradient is negative and proportionally increases the value in the next iteration. The vice versa is true for the predicted ˆhp larger than the ground truth. Integral Regression The gradient of the loss Lre (see Eq. 4) with respect to ˆhp can be estimated based on the chain rule as (see detailed derivation in Appendix A.6) hp ˆ hp = β hp |{z} 1 (value factor) (s( ˆJx Jx)(i ˆJx) + s( ˆJy Jy)(j ˆJy)) | {z } 2 (location factor) In Eq. 10, the gradient is split into two terms of interest: 1 and 2, which we name as the value factor and the location factor, respectively. The nature of the value and location factors gives strong indication to the learning process for integral regression methods. Firstly, the gradient p is proportional to the normalized predicted heatmap value hp as given in the value factor. This factor makes it prone to gradient vanishing wherever locations p have small heatmap predictions. Secondly, the location factor is a linear combination of i and j, and it manifests itself as a linear plane; far away points like corners are more likely to have large magnitudes. Though these two factors do not prevent the loss from decreasing during learning, they slow down training. We further elaborate this in Sec. 5.2. 5.2 A DETAILED ANALYSIS OF THE INTEGRAL REGRESSION HEATMAP We now conduct a detailed analysis of the gradients presented in Eq. 10 and show the typical cases that arise during learning. When referring to Ω, the domain of the heatmap, we use a coordinate system with the origin at the upper-left corner and increasing coordinates moving to the lower right corner (see Fig. 2 (a)). For simplicity of discussion, we assume that the ground truth joint Jgt is located somewhere in the lower-right quadrant of Ω, though the analysis holds for Jgt in all other quadrants of Ωanalogously. The four scenarios are visualized in Fig. 3. We base our analysis on the assumption that a larger magnitude of p has a larger influence on the weights of the network and causes a greater change at the location p in the next iteration, i.e. hn+1 p hn p γ n p, (11) where γ represents the step size or learning rate (see sketch of proof in Appendix A.7) and n denotes the update iteration. As the value and location factors have different contributions on the gradient from Eq. 10, and they are in turn affected by the size s and location µ of Φ, we can define four characteristic cases of Φ and outline how learning is affected in each case. (1) s is large, Φ has uniformly random values, e.g. a randomly initialized heatmap at the start of the training. As hp is in a similar range for all p, the value factor 1 (see Eq. 10) of all pixels are approximately similar. As such, the distinction between one pixel s gradient and that of another is determined by the location factor 2. Given that the ground truth coordinates (Jx, Jy) are in the lower-right quadrant, as long as (Jx, Jy) are both greater than the predicted coordinate ( ˆJx, ˆJy), p has a gradation that increases progressively towards the bottom right corner. Generally speaking, the location factor pushes the activations of the heatmap towards the corner of the correct quadrant as corners own the largest gradients in the resultant linear plane of the location factor 2. (2) s is small, µ is far from the ground truth J, e.g. when µ is in the upper left quadrant of Ωwhile the ground truth Jgt is at the bottom right. A typical scenario is if the heatmap activates around the Published as a conference paper at ICLR 2022 Figure 3: Toy example of regression update (left) and detection update (right). The black triangle denotes the prediction (expectation value for regression and max value for detection) while the red dot denotes the ground truth position. The arrow shows the movement of the prediction. left ankle for the ground truth right ankle. In this case, for the pixels outside of Φ, the value factor approaches zero, i.e. hp 0, and pushes p towards zero, hence the pixels receive very limited gradient updates. For pixels in Φ, the values decrease gradually until the value factor outside of Φ no longer dominates, i.e. they reach the same scale, at which point the heatmap returns to case (1). (3) s is small, µ is in the corner of the same quadrant as the ground truth Jgt. Initially, this case is similar to case (2). In the process of all the elements becoming the same scale, all the activations move towards the diagonal side and the prediction approaches the ground truth coordinate. (4) s is small, µ close to ground truth Jgt. This scenario occurs when the model is reasonably trained and can roughly localize the joint. The gradient at the ground truth pixel Jgt is always negative or zero, i.e. (Jx,Jy) = β h(Jx,Jy)( ˆJx Jx + ˆJy Jy ). (12) This non-positive value guides the network to predict a large heatmap response at the ground truth location, i.e. a large h(Jx,Jy); this in turn increases the gradient (Jx,Jy). This property makes the network more likely to predict a few exceptionally large pixels, thus shrinking the support region Φ and leading to a very small σ in Sec. 4.1. For detection-based methods, the ground truth heatmaps have already explicitly pointed out whether or not a pixel should be activated in the heatmap. Combined with its effective pixel-wise loss, the network learns faster than regression-based methods; this is validated in the next section. 5.3 EXPERIMENTAL VERIFICATION Idealized Sample We start by considering the case of one sample for a single joint and visualize the progression of a (64, 48) heatmap as it gets updated by Eq. 11. While n p can be calculated mathematically by Eq. 10, we use autograd in Pytorch to obtain the gradients, which we verify to be equivalent. The four cases from Sec. 5.2 are initialized randomly, with a symmetric Gaussian (σ = 2) centered at some upper left point, with a linear plane at the lower right quadrant, and a symmetric Gaussian (σ=2) centered at ground truth, respectively. The progression of the differently initialized heatmaps can be observed in Fig. 3. Mathematically, it is clear from Eq. 2 that the resulting joint coordinate location from integral regression can align with the ground truth joint even if the underlying heatmap does not follow a localized heatmap model. This is also illustrated in the first three cases discussed in Sec. 5.2. However, from the thousands of training samples observed over many epochs, we posit that during real-world training, the network can only consistently lower the loss if the heatmap H is learned to represent P(J|I). As such, the activation region Φ will be correctly localized over the corresponding semantic region in the image. Instead, what slows down convergence is the learning of Φ, or rather, the lack of direct guidance to yield a correct Φ as per detection methods. For the same four cases, we visualize the updated heatmap of Eq. 11 from the gradients defined by the detection loss in Eq. 9. Published as a conference paper at ICLR 2022 (a) Comparisons of detection and regression in real training (b) Training Efficiency of Detection versus Regression on COCO validation set Figure 4: (a) Comparisons of detection and regression in real training on an image selected from the COCO validation set at epoch 1 and 10. Results of detection have already roughly stabilized while regression can only localize to a small extent in epoch 10 with some background pixels still activated. We enlarge the right ankle part of regression in the red box for better visualization. More examples are shown in Appendix A.8. (b) The regression method converges slower than the detection method. Unlike regression, each case leads consistently to the same localized Φ centered on the ground truth joint coordinate, regardless of the initialization. Real World Samples The four cases listed in Sec. 5.2 are not so clearly observable in real-world training as the output is based on the optimization results of batch-wise training in a real network. To verify the results, we compare the training progression for a detection and an integral regression network under the experimental settings described in Sec. 4.1. Fig. 4 (b) compares the training speed of both detection and integral regression and shows the slower convergence of integral regression. We visualize for one sample from the MSCOCO validation set the heatmap at epochs 1 and 10 to compare the training progression for detection and integral regression. Fig. 4 (a) shows the heatmaps of the left eye and right ankle joints. For regression, after one epoch, the activations are spread over a quarter to half of the heatmap. This aligns roughly to a mix of Case 1 and 3 where activations are in the correct quadrant but not yet localized to a small region of support. After 10 epochs of training, predictions for the left eye (which is actually occluded, making it a hard sample) are localized with several activations of approximately the same value (all shown as red). For the right ankle (which is clearly present and is an easy sample), the network predicts heatmap values that have collapsed spatially and dominate at one or two pixel locations. In contrast, the detection method presents a well-localized heatmap after only one epoch of training. While there is some activation on the left ankle in epoch 1, this is unlikely to cause errors for the argmax decoding. After 10 epochs, all activations are centered around the correct right ankle. 6 CONCLUSION In this paper, we dive deeper into integral pose regression by comparing the differences in the heatmap activations and the gradients between detection and integral regression methods. Our theoretical analysis paired with empirical verification shows that the shrinkage of heatmaps of the integral regression method is the main cause for its lower performance compared to detection. Furthermore, we investigate the gradients of integral regression by giving theoretical evidence, toy problem and real world illustrations to show that the learning of integral regression is slower to converge as a result of only implicit supervisory signals. To mitigate the slow convergence and poor performance, we propose a simple spatial supervision on top of the existing framework. Acknowledgements This research / project is supported by the Ministry of Education, Singapore, under its MOE Academic Research Fund Tier 2 (STEM RIE2025 MOE-T2EP20220-0015). Published as a conference paper at ICLR 2022 COCO Leader Board. https://cocodataset.org/, 2020. Zhen-Hua Feng, Josef Kittler, Muhammad Awais, Patrik Huber, and Xiao-Jun Wu. Wing loss for robust facial landmark localisation with convolutional neural networks. In CVPR, pp. 2235 2245, 2018. Kerui Gu, Linlin Yang, and Angela Yao. Removing the bias of integral pose regression. In ICCV, pp. 11067 11076, 2021. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, pp. 770 778, 2016. Junjie Huang, Zheng Zhu, Feng Guo, and Guan Huang. The devil is in the details: Delving into unbiased data processing for human pose estimation. In CVPR, pp. 5700 5709, 2020. Umar Iqbal, Pavlo Molchanov, Thomas Breuel Juergen Gall, and Jan Kautz. Hand pose estimation via latent 2.5 d heatmap regression. In ECCV, pp. 118 134, 2018. Wenbo Li, Zhicheng Wang, Binyi Yin, Qixiang Peng, Yuming Du, Tianzi Xiao, Gang Yu, Hongtao Lu, Yichen Wei, and Jian Sun. Rethinking on multi-stage networks for human pose estimation. ar Xiv preprint ar Xiv:1901.00148, 2019. Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, pp. 740 755. Springer, 2014. Alejandro Newell, Kaiyu Yang, and Jia Deng. Stacked hourglass networks for human pose estimation. In ECCV, pp. 483 499. Springer, 2016. Aiden Nibali, Zhen He, Stuart Morgan, and Luke Prendergast. Numerical coordinate regression with convolutional neural networks. ar Xiv preprint ar Xiv:1801.07372, 2018. Matteo Ruggero Ronchi and Pietro Perona. Benchmarking and error diagnosis in multi-instance pose estimation. In ICCV, pp. 369 378, 2017. Adrian Spurr, Umar Iqbal, Pavlo Molchanov, Otmar Hilliges, and Jan Kautz. Weakly supervised 3d hand pose estimation via biomechanical constraints. In ECCV, pp. 211 228, 2020. Ke Sun, Bin Xiao, Dong Liu, and Jingdong Wang. Deep high-resolution representation learning for human pose estimation. In CVPR, pp. 5693 5703, 2019. Xiao Sun, Bin Xiao, Fangyin Wei, Shuang Liang, and Yichen Wei. Integral human pose regression. In ECCV, pp. 529 545, 2018. Jonathan Tompson, Arjun Jain, Yann Le Cun, and Christoph Bregler. Joint training of a convolutional network and a graphical model for human pose estimation. Neur IPS, 27, 2014. Bin Xiao, Haiping Wu, and Yichen Wei. Simple baselines for human pose estimation and tracking. In ECCV, pp. 466 481, 2018. Linlin Yang, Shicheng Chen, and Angela Yao. Semihand: Semi-supervised hand pose estimation with consistency. In ICCV, pp. 11364 11373, 2021. Feng Zhang, Xiatian Zhu, Hanbin Dai, Mao Ye, and Ce Zhu. Distribution-aware coordinate representation for human pose estimation. In CVPR, pp. 7093 7102, 2020. Beier Zhu, Chunze Lin, Quan Wang, Renjie Liao, and Chen Qian. Fast and accurate: Structure coherence component for face alignment. ar Xiv preprint ar Xiv:2006.11697, 2020. Published as a conference paper at ICLR 2022 A.1 EXTENSION OF FIG. 2 (B) We provide an extended plot of Fig. 2 (b), which includes both detection and regression on easy, medium, and hard cases in Fig. I. A.2 PROOF FOR EQ. 7 EQ. 8 IN SEC. 4.2 This subsection gives the detailed proof for Eq. 7>Eq. 8, which means p Φ(µ,s) w(p)||Jgt p||2 ||Jgt µ||2, where X p Φ(µ,s) w(p) = 1. (I) In this way, this equation can be rewritten as p Φ(µ,s) w(p)||Jgt p||2 X p Φ(µ,s) w(p)||Jgt µ||2. (II) As the distribution of Φ(µ, s) is radially symmetric, a sufficient but unnecessary proof is that for any symmetric pair {p1(µx a, µy b), p2(µx + a, µy + b)}, it satisfies (x µx a)2 + (y µy b)2 + p (x µx + a)2 + (y µy + b)2 2 p (x µx)2 + (y µy)2. (III) According to Triangle Inequality in normed vector space, we arrive at ||u||2 + ||v||2 ||u + v||2. (IV) When we set u = (x µx a, y µy b) and v = (x µx + a, y µy + b), Eq. II is established. A.3 PEARSON CHI-SQUARE VALUES ON DIFFERENT CASES We report the detailed Pearson Chi-square values between heatmaps of both detection and regression in different scenarios and Gaussian templates with different values of σ in Table. I, II, III. We set the threshold of A(s) > 0.8 to show that the support region can represent the whole heatmaps. Therefore, for easy and medium cases, s=4; while for hard cases, s=8. 2 4 6 8 10 12 14 s Activation Sum A(s) Regression-Easy Regression-Medium Regression-Hard Regression-Overall Detection-Easy Detection-Medium Detection-Hard Detection-Overall Figure I: Activation Sum A(s) of both detection and regression method on easy, medium, hard, and overall cases. Published as a conference paper at ICLR 2022 s σ regression detection 2 3 4 5 2 3 4 5 8 >1000 4.76 4.12 3.95 >1000 3.51 3.31 3.25 Table I: Results of the Pearson Chi-square test between distribution of heatmap H and Gaussian templates T with different window size s and standard deviation σ in hard cases. σ =5 is the bestmatching template for both detection and regression. σ stops at 5 as the Gaussian templates with size 8 become similar after. s σ regression detection 1 2 3 4 1 2 3 4 4 >1000 3.25 3.65 3.73 >1000 1.76 0.76 0.77 Table II: Results of the Pearson Chi-square test between distribution of heatmap H and Gaussian templates T with different window size s and standard deviation σ in medium cases. For detection, best matched σ should be between 3 and 4; while for regression, optimal σ in medium case should be between 1 and 2. A.4 DETAILS OF COMBINING PRIOR LOSS AND INTEGRAL LOSS This subsection demonstrates the influence of weights of prior loss and integral loss, and the magnitudes of two gradients regarding the shared heatmaps during the training. Let us assume wp and wi are weights of the prior and integral losses, respectively. We consider the ratio wp Lp : wi λLi, where λ = 10 2, to set the two to approximately equal magnitudes at the initial epoch. Table V shows that the AP heavily depends on the value of σ but for some fixed σ (within each row), the ratio of the loss weightings has less impact and the results are stable within 1 2%, with a larger influence only at the extreme settings of 100 : 1 or 1 : 100. Table IV (b) provides the ratios of the gradient magnitudes to help explain why the weightings have little impact. Table IV (b) gives the mean ratio gp gi where gp and gi are the gradient magnitudes of the prior loss and integral loss, respectively, over the course of training. A column of 2/1/0/-1/-2 means 102/101/100/10 1/10 2 ratio in gradients over the epochs. The table shows that regardless of the weighting, the gradient ratio decreases over the course of training (i.e. dropoff with each column). Apart from an extremely high weighting of the prior, i.e. 100:1 in column 1, the prior influences the training only in the initial 10 epochs. After epoch 10, the gradient ratio is around the same scale or smaller. A.5 DETAILS OF HEATMAP SHRINKAGE To find out whether the added prior influences the spread of the heatmaps, we investigate the shrinkage change of heatmaps by separating the samples. We separate the samples based on easy/medium versus hard cases, where the former generally produce localized heatmaps while the latter have dispersed heatmaps as per Fig. I; the alternative is to separate by thresholding the activation sum in a fixed area (Eq. 6). We consider the activation sum A(1) > 0.34 of the integral regression baseline as localized, otherwise as dispersed (threshold 0.34 selected based on Fig. I). We tally the activation sums for the localized and dispersed heatmaps, without and with the prior for different values of σ. Activation sums A(s)e/m, A(s)h, A(s)l, A(s)h, and Average Precision APe/m, APh, APl, APh indicate the extent of localization and the performance of the two conditions, respectively. An increase in activation sum indicates further heatmap localization; a decrease indicates dispersion. The No prior row is the baseline integral pose regression. From Table IV (a), both separations present the same trend. For localized cases, forcing σ to be degenerately small or too large (σ = 0.5 or σ = 4) increases/decreases the activation sum over the baseline, thereby confirming the effect of the prior as intended. The change in AP follows, i.e. decreases when σ becomes incompatible from being too small or too large. For dispersed cases, the prior shrinks the activation sum as expected. Like the localized case, there is an optimal range for which improvement in AP is better; degenerate σ will decrease the AP from the baseline. Published as a conference paper at ICLR 2022 s σ regression detection 0.5 1 2 3 4 1 2 2.5 3 3.5 4 4 >1000 3.24 3.53 3.75 3.89 >1000 2.52 0.64 0.72 0.85 0.88 Table III: Results of Pearson Chi-square test between distribution of heatmap H and Gaussian templates T with different window size s and standard deviation σ in easy cases. Accordingly, for detection, σ should be approximately 2.5; while for regression, optimal σ should be between 0.5 and 1. wp : λwi 100:1 10:1 1:1 1:10 1:100 +σ=0.5 63.6 64.8 65.1 65.7 66.4 +σ=1 67.6 68.1 68.5 68.3 67.8 +σ=2 68.5 70.8 71.4 70.6 67.8 +σ=3 68.2 69.4 69.8 69.2 68.1 +σ=4 64.2 66.1 66.9 66.9 67.0 wp : λwi 100:1 10:1 1:1 1:10 1:100 epoch 1 2 2 2 1 1 epoch 10 2 1 0 0 0 epoch 30 2 0 0 0 -1 epoch 50 1 -1 -1 -1 -1 epoch 80 0 -1 -1 -1 -2 (a) (b) Table IV: (a) Performance of the integral model adding different priors with different weight ratios wp : λwi. The performance depends largely on different σs, but less influenced with reasonable change. (b) Ratio of the gradient magnitudes of combined model (integral+σ = 2) with different weight ratios wp : λwi. The gradient ratio decreases over the course of training. After epoch 10, the gradient ratio is around the same scale or smaller. model A(2)e/m APe/m A(2)e/m APl A(2)h APh A(2)l APl No prior 0.65 70.4 0.71 72.3 0.35 44.7 0.33 38.9 +σ=0.5 0.75 68.2 0.79 69.8 0.36 43.9 0.36 35.2 +σ=1 0.63 71.9 0.70 73.9 0.36 45.2 0.31 39.1 +σ=2 0.54 74.2 0.59 76.1 0.41 47.4 0.37 39.8 +σ=3 0.49 73.0 0.56 74.9 0.39 47.6 0.36 39.3 +σ=4 0.45 70.2 0.55 71.7 0.39 46.9 0.38 39.5 Table V: Localization extent and performance of divided localized samples versus dispersed samples. The extent of localization largely depends on the added prior. With an optimal σ = 2 chosen, the previously localized heatmaps have been dispersed with improved performance. A.6 DERIVATION OF EQ. 10 Eq. 10 is obtained by the chain rule. Therefore, we list the value of each term as follows: ˆJ = s( ˆJx Jx) s( ˆJy Jy) , ˆJ hp = (i + j), hp ˆ hq = β hq(1 hq), q = p hp hq, q = p , (V) where q = (u, v) can be any pixel in Ω. Combining the terms, gradients at p can be calculated by ˆhp = β(( ˆJx Jx)i + ( ˆJy Jy)j) hp(1 hp) X β(( ˆJx Jx)u + ( ˆJy Jy)v) hq hp. (VI) Eq. 10 can be obtained by rearranging the terms of Eq. VI. A.7 DERIVATION OF EQ. 11 Let us suppose Hn = g(wn|I) where Hn is the heatmap of a given image I at the n-th iteration, g(wn) is the function of generating the heatmap, i.e. the backbone network, with weights at the n-th iteration. According to a Taylor expansion, we arrive at Published as a conference paper at ICLR 2022 Image Nose Left eye Right eye Left ear Right ear Left shoulder Right shoulder Right elbow wrist Left hip Right hip Right ankle Figure II: Visualization of the training performance of integral pose regression on one single image at epoch 0, 5, 10, 15, 20 (from top to down). Corner or edge pixels of the correct quadrant are largely activated in the epoch 0 (highlighted in the yellow boxes). Further training is focused on narrowing down the activated region (shown in the red boxes). Image Nose Left eye Right eye Left ear Right ear Left shoulder Right shoulder Right elbow wrist Left hip Right hip Right ankle Figure III: Visualization of the training performance of detection-based methods on one single image at epoch 0, 5, 10, 15, 20 (from top to down). One epoch has already enabled the network to roughly localize the joint compared with regression-based method (in the yellow boxes). Further training of detection-based methods makes the predictions more stable and precise (highlighted in red boxes). g(wn+1) = g(wn) + g (wn) 1! (wn+1 wn) + g (wn) 2! (wn+1 wn)2 + . . . , (VII) where wn+1 wn= α g(w)= α L Hg (wn) according to gradient descent, and (wn+1 wn)i is negligible compared with (wn+1 wn) when i>1. Therefore, Eq. VII can be rewritten as Hn+1 = g(wn+1) g(wn) + g (wn)(wn+1 wn) = Hn α L H||g (wn)||2. (VIII) Therefore, for each pixel hp in H, the update rule is hn+1 p hn p γ n p, (IX) where γ represents α g (wn) , hn p and n p denote heatmap and gradient value located at p in the heatmap at the nth iteration. h0 p denotes the original heatmap. Published as a conference paper at ICLR 2022 A.8 REALWORLD VISUALIZATIONS We show the visualization of the training performance of integral pose regression and detectionbased methods on one single image in Figs. II and III, respectively. We can see that the further training of integral pose regression is focused on narrowing down the activated region while the further training of detection-based methods makes the predictions more stable and precise.