# generative_point_cloud_registration__4aae9f0f.pdf Generative Point Cloud Registration Haobo Jiang 1 Jin Xie 2 Jian Yang 3 Liang Yu 4 Jianmin Zheng 1 Abstract In this paper, we propose a novel 3D registration paradigm, Generative Point Cloud Registration, which bridges advanced 2D generative models with 3D matching tasks to enhance registration performance. Our key idea is to generate cross-view consistent image pairs that are wellaligned with the source and target point clouds, enabling geometry-color feature fusion to facilitate robust matching. To ensure high-quality matching, the generated image pair should feature both 2D-3D geometric consistency and crossview texture consistency. To achieve this, we introduce Match-Control Net, a matching-specific, controllable 2D generative model. Specifically, it leverages the depth-conditioned generation capability of Control Net to produce images that are geometrically aligned with depth maps derived from point clouds, ensuring 2D-3D geometric consistency. Additionally, by incorporating a coupled conditional denoising scheme and coupled prompt guidance, Match-Control Net further promotes cross-view feature interaction, guiding texture consistency generation. Our generative 3D registration paradigm is general and could be seamlessly integrated into various registration methods to enhance their performance. Extensive experiments on 3DMatch and Scan Net datasets verify the effectiveness of our approach. [Code] 1. Introduction Point cloud registration is a problem of finding the optimal rigid transformation, comprising a 3D rotation and a 3D translation, which aligns the source and target point clouds precisely. It plays an important role in various downstream computer vision applications, such as 3D reconstruction, Li DAR SLAM, and object localization. However, real-world 1Nanyang Technological University, Singapore 2Nanjing University, China 3Nankai University, China 4Alibaba Group, China. Correspondence to: Jianmin Zheng . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Conventional 3D Registration Generative 3D Registration Geometry-only Matching & Pose Estimation Match-Control Net Color-Geometric Matching & Pose Estimation Generated Images Figure 1. Paradigm comparison of our generative point cloud registration with conventional methods. Unlike geometry-only matching in previous methods, our approach introduces Match Control Net, a matching-specific 2D generative model that generates cross-view images pairs from point cloud data, providing rich color cues for enhanced geometric matching and pose estimation. challenges like low overlap and noisy points still hinder its adoption in broader real-world scenarios. Existing 3D registration methods can be roughly categorized into traditional approaches and data-driven deep methods. The traditional approaches include optimization-based fine alignment methods (Besl & Mc Kay, 1992; Yang et al., 2013), which iteratively perform least-squares pose optimization for precise alignment, and handcrafted descriptorbased coarse alignment methods (Rusu et al., 2008; 2009), which capture local geometry to establish correspondences for hypothesize-and-verify registration. Deep registration methods, whether end-to-end (Yew & Lee, 2020; 2022) or descriptor-based (Huang et al., 2021; Qin et al., 2022; Jiang et al., 2023a), exploit the power of deep neural networks to learn discriminative deep 3D features for robust matching. These deep methods significantly enhance the quality of estimated correspondences and improve registration accuracy. Despite the impressive performance achieved by current point cloud registration methods, their robustness remains limited in challenging scenarios that contain low overlap, repetitive patterns, or noisy points. Recent RGB-D registration studies (Yuan et al., 2023; Mu et al., 2024) have shown that incorporating rich texture and semantic cues from RGB images would significantly enhance the distinctiveness of point cloud descriptors, leading to improved matching accuracy. However, in geometry-only point cloud registration, the RGB images corresponding to the point clouds are unavailable, and existing methods rely solely on 3D geometric information for correspondence estimation and pose calcu- Generative Point Cloud Registration lation. This raises an interesting question: Can we still leverage color information to enhance geometry-only point descriptors for enhanced 3D registration? Motivated by this question and inspired by the recent successes of generative AI models (Ho et al., 2020; Song et al., 2020; Yang et al., 2023; Li et al., 2025; Rombach et al., 2022; Zhang et al., 2023; Jiang et al., 2023b; Wang et al., 2024), we introduce Generative Point Cloud Registration, a new 3D matching paradigm that bridges the task gap between the 2D generative models and 3D matching tasks to enhance registration performance. Our key idea is to generate cross-view image pairs that are well-aligned with the corresponding source and target point clouds. These images provide rich color information to complement geometric features, enabling more robust matching (see Fig. 1). Unlike prevalent 2D generative models that focus on single image generation, our matching-specific image generation is pairwise. Importantly, to ensure high matching quality, the generated cross-view image pair should feature two key properties: 2D-3D geometric consistency and cross-view texture consistency. To achieve this, we introduce Match Control Net, a matching-specific, controllable 2D generative model. Match-Control Net leverages Control Net s depthconditioned generation capabilities to produce images geometrically aligned with depth maps (derived from the point cloud pairs), ensuring 2D-3D geometric consistency. Additionally, by incorporating coupled conditional denoising and coupled prompt guidance, Match-Control Net enables effective cross-view image feature interaction, achieving mutual texture message passing and thereby enhancing crossview texture consistency. Finally, we propose a zero-shot geometric-color fusion mechanism that leverages pretrained large vision models (e.g., DINOv2 (Oquab et al., 2023) and Stable Diffusion (Rombach et al., 2022)) to extract discriminative zero-shot features of generated images for enhancing geometric descriptors via weighted concatenation. It should be pointed out that our Generative Point Cloud Registration framework can operate in both zero-shot and few-shot settings (with minimal fine-tuning samples), each providing valuable color information to enhance precision. Moreover, our framework is general and can be integrated with various 3D registration methods to enhance their matching accuracy. Experiments on 3DMatch and Scan Net datasets validate the effectiveness of our proposed method. To summarize, our contributions are as follows: We propose a new Generative Point Cloud Registration paradigm, aimed at generating cross-view image pairs for both source and target point clouds, thereby providing rich color information for effective geometric-color feature fusion and improved matching quality. Unlike conventional single-image generation, we develop an effective Match-Control Net for matching- specific, pairwise image generation. It incorporates depth-conditioned generation, coupled conditional denoising, and coupled prompt guidance to ensure that the generated image pairs maintain 2D-3D geometric consistency and cross-view texture consistency. Our Generative Point Cloud Registration framework is general and plug-and-play. Benefiting from our effective zero-shot geometric-color feature fusion and XYZ-RGB fusion schemes, it can be integrated with various 3D registration approaches to provide free-lunch color information, enhancing their performance. 2. Related Work Traditional 3D Registration Methods. Traditional point cloud registration methods are typically categorized into coarse and fine registration approaches. Iterative Closest Point (ICP) (Besl & Mc Kay, 1992), a prominent fine registration method, iteratively computes nearest-neighbor correspondences and performs least-squares optimization for pose estimation. Go-ICP (Yang et al., 2013) enhances ICP s robustness to initialization errors through a branch-andbound (Bn B) global search. Trimmed ICP (Chetverikov et al., 2002) further improves robustness by optimizing over minimal subsets to handle outliers. Additional variants like (Sharp et al., 2002; Fitzgibbon, 2003; Bae & Lichti, 2008; Gressin et al., 2013; Deng et al., 2018) also demonstrate promising precision in fine alignment. Coarse registration methods generally combine handcrafted geometric descriptors with robust pose estimators, such as RANSAC. (Johnson & Hebert, 1999) develops the spin image-based shape descriptors for surface matching and object recognition. USC (Tombari et al., 2010) improves the feature descriptors using an shape context-aware unique local reference frame to improve matching accuracy. SHOT (Salti et al., 2014) introduces a 3D histogram-based feature using normal vectors to describe surface. PFH (Rusu et al., 2008) and FPFH (Rusu et al., 2009) constructs a discriminative and efficient local descriptor based on the oriented histogram with pairwise 3D representations. Other notable coarse methods, including (Mohamad et al., 2014; 2015; Xu et al., 2019; Huang et al., 2017; Ge, 2017), have also achieved impressive registration precision. Learning-based Deep Registration Methods. Deep registration methods primarily consist of end-to-end approaches and deep descriptor-based methods. For end-to-end approaches, DCP (Wang & Solomon, 2019) introduces differentiable soft correspondences for SVD-based pose estimation. RPM-Net (Yew & Lee, 2020) incorporates the Sinkhorn layer and an annealing strategy to mitigate outlier inference. CEMNet and Latent CEM (Jiang et al., 2021a;b) formulate the 3D registration task as a Markov decision process, and introduce a reinforcement learning-driven learning Generative Point Cloud Registration Color Point Cloud Registration (e.g., Color PCR) 1. Zero-shot Geometric-Color Feature Fusion Pretrained Large Vision Models (e.g., DINOv2, SD ) Geometric Descriptors (e.g., FCGF, Predator ) 2. XYZ-RGB Fusion Match-Control Net Interaction Couple Decouple Correspondence + Pose Estimation Cross-view Image Generation Figure 2. Pipeline of Generative Point Cloud Registration. Given a source and a target point cloud, we first apply Match-Control Net to generate their corresponding images. Next, we employ either zero-shot geometric-color feature fusion or XYZ-RGB fusion to create color-enhanced geometric descriptors, enabling high-quality correspondence estimation and robust pose estimation. framework for planning-based trial-and-error pose searching (Liu et al., 2024; 2023; Jiang et al., 2021c; 2022). Reg TR (Yew & Lee, 2022) designs an effective transformerbased correspondence regression module, addressing largescale indoor scene registration in an end-to-end manner. For deep descriptor-based methods, 3DMatch (Zeng et al., 2017) employs a Siamese 3D CNN to extract local geometric features for patchwise matching. FCGF (Choy et al., 2019) develops a fully convolutional network to learn dense 3D features for pointwise matching. Predator (Huang et al., 2021) introduces a cross-attention transformer between point cloud pairs for overlap perception and robust registration. Geo Transformer (Qin et al., 2022) integrates geometric embeddings into the transformer, enhancing feature discrimination. Ro ITr (Yu et al., 2023) designs a rotation-invariant transformation to further improve the rotational robustness of geometric descriptors. Other methods (Wang et al., 2023; Bai et al., 2020; Li & Harada, 2022; Li et al., 2020; Choy et al., 2020; Chen et al., 2023; Fu et al., 2021; Ao et al., 2023) also demonstrate impressive performance in 3D registration. Beyond traditional and learning-based frameworks, this work introduces a new paradigm: Generative Point Cloud Registration. By integrating advanced 2D generative models with the 3D registration domain, our approach generates complementary color information for input point cloud pairs, producing color-enhanced geometric descriptors to improve precision. 3. Approach Problem Setting. Given a pair of source and target point clouds P = {pi R3 | i = 1, . . . , N} and Q = {qi R3 | i = 1, . . . , M}, point cloud registration seeks to recover their rigid transformation T = {R, t} SE(3), comprising a rotation R SO(3) and a translation t R3, to align them precisely. The optimal rigid transformation is typically computed by solving: (p ,q ) C R p + t q 2 2 , (1) where C denotes the ground-truth correspondences between source and target point clouds. However, C is generally unknown in practical usage, and we need estimate a set of putative correspondences through finding feature nearest neighbor among the pointwise geometric descriptors. 3.1. Motivation Recent RGB-D point cloud registration methods (Yuan et al., 2023; Mu et al., 2024) have demonstrated that RGB images can significantly enhance geometry-only descriptors by providing rich color and semantic information. This enhancement facilitates the construction of higher-quality correspondences, leading to more robust registration. However, in the context of 3D matching tasks, we focus on geometry-only registration using pure point clouds, as the relevant RGB data is unavailable. To overcome this limitation, we introduce Generative Point Cloud Registration, a general framework designed to generate high-quality RGB data for both point clouds, enabling geometric-color feature fusion for enhanced matching. Notably, unlike the conventional single-image generation focused by prevalent 2D generative models (Rombach et al., 2022; Zhang et al., 2023), our matching-specific image generation is pairwise, which should satisfy two key criteria: (i) 2D-3D Geometric Consistency: The generated images should preserve the geometric structure and spatial layout of their respective point clouds to ensure accurate pixel-topoint correspondences and avoid introducing noise; (ii) Cross-view Texture Consistency: The generated image pair should maintain consistent textures for correspondences. Otherwise, inconsistent textures would reduce feature similarity of correspondences, leading to mismatches. 3.2. Zero-Shot Geometric Consistency Generation In this section, we first address the 2D-3D geometric consistency generation through Control Net (Zhang et al., 2023), a variant of Stable Diffusion (Rombach et al., 2022) with spatially localized image conditions (e.g. Canny edge maps and depth maps). Stable Diffusion. Stable Diffusion is a widely used latent diffusion model for text-to-image generation. It operates within the latent space of a pretrained autoencoder, where a denoiser ϵθ(xt; t, c) (conditioned on the timstamp t and tokenized text prompt c) gradually refines the noisy latent Generative Point Cloud Registration Match-Control Net Feature Interaction Couple Decouple Control Net Control Net Independent Generation w/o Feature Interaction Consistent Geometry Consistent Texture Consistent Geometry Consistent Texture Figure 3. Instead of independently performing Control Net to generate source and target images, our Match-Control Net integrates their denoising generation processes into a unified framework, facilitating feature interaction (i.e., mutual texture message passing) and enhancing their cross-view texture consistency. feature xt to clean one for image decoding. The denoiser follows a UNet architecture with an encoder, middle block, and skip-connected decoder, incorporating stacked transformer and residual modules. Each transformer module utilizes intra-image self-attention for contextual understanding and prompt-to-image cross-attention to guide generation. Control Net-driven 2D-3D Geometric Consistency. Control Net further equips the denoiser of Stable Diffusion with a learnable encoder copy for encoding the conditional image c I, forming a conditional denoiser: ϵθ(xt; t, c, c I). Consequently, the encoded condition features is concatenated with the original encodings of noisy latent representations xt for conditional feature decoding via skip connections. Notably, Control Net allows the use of depth maps as conditional inputs to generate RGB images that preserve geometric structures well-aligned with the provided depth prior. This capability perfectly aligns with our objective and motivates us to convert the source and target point clouds into their corresponding depth maps, DP and DQ RH W 1, via the intrinsic matrix. Then, each produced depth map can be independently used to condition Control Net to produce the geometrically consistent image. 3.3. Zero-Shot Texture Consistency Generation Although the original Control Net can generate source and target image pairs that are geometrically well-aligned with the given point cloud pair, the texture details of the corresponding regions between the generated image pair often differ as shown in Fig. 3 (left). This texture inconsistency primarily arises because the denoising processes for the source and target images operate independently as follows: ϵθ(x P t ; t, c, d P) x P t 1, ϵθ(x Q t ; t, c, d Q) x Q t 1, (2) with each unaware of the colors produced by the other. Here, x P t , x Q t RH W d denote the noisy latent representations corresponding to source and target images; d P, d Q RH W d represent the encoded features of depth maps DP and DQ via optimized zero convolutions of Control Net. This insight motivates us to combine source and target image denoising generation processes into a joint denoising pass, thereby enabling mutual texture message passing and promoting texture consistency (see Fig. 3). Based on this motivation, we establish Match-Control Net, an improved Control Net variant for matching-specific, conditional image generation. Following Sec. 3.2, we still take depth maps derived by point clouds as conditional images so as to inherit Control Net s 2D-3D geometric consistency generation capability. Additionally, we introduce two key designs: coupled conditional denoising and coupled prompt guidance to achieve the cross-view texture consistency generation. The details of these two designs are as follows: Coupled Conditional Denoising. To achieve mutual texture message passing, a straightforward approach is to build two denoisers and incorporate an additional cross-denoiser attention module to facilitate their message passing. However, running two denoisers simultaneously is inefficient, and such significant architectural changes would require extensive model fine-tuning. To enable effective cross-view message passing without any finetuning (i.e., zero-shot), we propose an efficient coupled conditional denoising scheme for joint, interactive source and target image generations. Specifically, we expand the noisy latent representation x P(Q) t with shape [H , W , d] to a coupled one x PQ t with the extended shape [2H , W , d]. Also, we vertically concatenate the source and target depth maps into a coupled one DPQ R2H W 1, and further emply the Control Net s zero convolutions to encode it into the condition features d PQ R2H W d. Consequently, without any architectural modifications or parameter finetuning, the original conditional denoiser can be directly employed for our coupled conditional denoising: ϵθ x PQ t ; t, c, d PQ x PQ t 1, (3) forming the denoising Markov chain: x PQ T x PQ 1 x PQ 0 . Here, the initial coupled latent representation x PQ T is sampled from a standard Gaussian distribution N(0, I). To further clarify the cross-view texture message passing during our coupled denoising process, we formulate the selfattention mechanism within the denoiser as SA(x PQ t ) = softmax((x PQ t Wq)(x PQ t Wk) d )(x PQ t Wv), (4) where Wq, Wk and Wv Rd d denote the projection matrices for queries, keys and values, respectively. Eq. 4 illustrates that by coupling the source and target noisy latent representations, each feature element can establish longrange dependencies with all feature elements from both the source and target feature maps, allowing effective crossview feature interaction and texture-aware message passing for promoting texture consistency generation. Generative Point Cloud Registration Coupled Prompt Guidance. Although the aforementioned coupled denoising generative mechanism has provided the essential components for texture consistency generation, the denoiser still fails to produce the image pair with expected cross-view texture consistency. The core reason is that the denoiser does not know what kind of image the user expects to generate, and we need to tell it what to do. Our important finding is that when we use a specific prompt (named coupled prompt) as below to tell our Match-Control Net to produce the vertically stacked images with consistent layout and elements: Generate two vertically stacked images that are captured from the different viewpoints in a same scene. The images should feature the same environment whether indoor or outdoor, like a living room, office, street, or natural landscape with very subtle differences between them. Overall, the layout and key elements remain the same. , the denoiser can naturally be guided to recover consistent textures without any model fine-tuning. To the best of our knowledge, we are the first to uncover and utilize this inherent capability of the pre-trained Control Net for zero-shot pairwise image generation enjoying both 2D-3D geometric consistency and cross-view texture consistency. 3.4. Few-Shot Consistency Fine-tuning Although our zero-shot Match-Control Net above has demonstrated promising pairwise consistency generation capabilities, Fig. 4 shows that some corresponding regions may still exhibit geometric or texture inconsistency issues. To mitigate it, we further propose a few-shot finetuning mechanism to improve the consistency generation quality of our Match Control Net. It s noted that we only finetune the learnable encoder copy of Match-Control Net rather than the all parameters to preserve the powerful generation ability of Stable Diffusion. Specifically, we first collect a set of RGB-depth pairs, {((IP, DP), (IQ, DQ))j} (j denotes the sample index), as the training data. For each sample pair, we then concatenate its depth maps and RGB images into coupled ones {(IPQ, DPQ)j}. Finally, we use the loss function below to finetune the denoiser: L = Ex PQ t ,t, c,d PQ,ϵ N(0,1) h ϵ ϵθ(x PQ t , t, c, d PQ) 2 2 i , (5) where x PQ t represents the diffused latent representation of the coupled image IPQ; c denotes the token sequence of our coupled prompt. Our experiments show that even with a limited number of samples ( 3K), few-shot finetuning can effectively improves the quality of consistency generation. 3.5. Geometric-Color Fused Point Descriptor In this section, we focus on how to enhance the geometric representations of point clouds with the free-lunch color Zero-shot Finetuned Figure 4. Compared to the zero-shot Match-Control Net (top), the finetuned Match-Control Net can tend to achieve higher 2D-3D geometric consistency and the cross-view texture consistency. information from generated source and target images, IP and IQ. We provide two geometric-color fusion schemes: Zero-Shot Geometric-Color Feature Fusion. Inspired by the powerful RGB representations of large vision models, we utilize them to directly extract zero-shot semantic features from the generated images. Specifically, we employ two widely-used vision foundation models: DINOv2 (Oquab et al., 2023) and Stable Diffusion (Rombach et al., 2022) for image encoding, achieving corresponding feature maps. These feature maps are then projected into the point cloud space using the camera intrinsic matrix, yielding pointwise color descriptors: {f rgb pi } and {f rgb qi } for both source and target point clouds. Finally, we apply a simple weighted concatenation to combine the RGB descriptors with the geometric descriptors, resulting in fused descriptors as follows: fpi = [ω f geo pi ; (1 ω) PCA(f rgb pi )] Rdgeo+drgb, (6) where [ ; ] denotes the feature concatenation operator; ω is the fusion weight and f geo pi represents the geometric descriptors; PCA( ) denotes the principal component analysis function to compress the feature dimension of the color descriptor to fit that of the geometric descriptor. The same fusion scheme is also applied to the target point clouds. Notably, this zero-shot geometric-color fusion approach is general and can be applied to a variety of geometric descriptors, whether traditional or deep descriptors. XYZ-RGB Fusion. This fusion scheme directly projects the generated source and target RGB images into the point cloud space. The resulting RGB values are then concatenated with the point coordinates of the point clouds, forming 6D color source and target point clouds, as shown in Fig. 5 (left). These color point clouds are subsequently used as inputs to the color point cloud registration method, like Color PCR (Mu et al., 2024), for 3D registration. 4. Experiments 4.1. Experimental Setting Implementation Details. During the few-shot fine-tuning stage, we randomly select 3,000 sample pairs from the Scan- Generative Point Cloud Registration Table 1. Comparison of the methods on rotation, translation, and Chamfer distance on Scan Net (Dai et al., 2017) benchmark dataset. Rotation (deg) Translation (cm) Chamfer (mm) Accuracy Error Accuracy Error Accuracy Error Methods 5 10 45 Mean Med. 5 10 25 Mean Med. 1 5 10 Mean Med. FPFH (Rusu et al., 2009) 41.4 56.7 73.3 39.2 7.1 17.5 35.1 50.9 79.5 23.5 32.3 48.0 53.0 159.6 6.5 Lepard (Li & Harada, 2022) 63.3 75.5 84.1 24.9 3.3 31.3 56.4 72.3 48.4 8.1 51.6 69.1 73.0 89.2 0.9 Reg TR (Yew & Lee, 2022) 72.5 83.8 94.1 10.2 2.3 44.3 65.6 80.0 27.7 5.8 61.0 76.6 80.9 54.0 0.5 Ro ITr (Yu et al., 2023) 70.0 77.5 83.7 24.1 2.3 40.3 62.3 75.1 45.6 6.5 58.8 72.4 75.4 94.1 0.6 FCGF (Choy et al., 2019) 78.9 84.2 87.5 19.4 1.5 55.3 70.7 79.7 37.8 4.3 67.3 78.2 80.3 100.7 0.4 Generative FCGFDINOv2 81.0 86.2 89.4 16.5 1.4 57.3 72.6 80.9 33.9 4.0 68.9 79.5 81.5 97.4 0.3 Generative FCGFSD 82.9 90.0 94.4 8.4 1.6 56.4 73.0 82.7 21.7 4.1 67.7 80.9 83.7 66.0 0.4 Improvement 4.0 5.8 6.9 11.0 0.1 2.0 2.3 3.0 16.1 0.3 1.6 2.7 3.4 34.7 0.1 Predator (Huang et al., 2021) 64.3 75.2 82.6 26.3 3.2 30.1 54.8 69.2 48.7 8.4 50.8 66.9 70.6 93.2 1.0 Generative Predator DINOv2 67.0 78.0 87.2 19.0 3.0 30.7 56.0 70.3 41.4 8.1 52.0 67.8 71.3 79.2 0.9 Generative Predator SD 70.7 81.3 88.7 17.0 2.8 33.0 59.4 73.3 36.6 7.5 54.7 70.8 74.2 72.5 0.8 Improvement 6.4 6.1 6.1 9.3 0.4 2.9 4.6 4.1 12.1 0.9 3.9 3.9 3.6 20.7 0.2 Geo Trans (Qin et al., 2022) 71.5 78.0 83.4 26.2 2.0 48.4 65.2 74.6 51.9 5.2 62.0 72.5 75.0 97.3 0.5 Generative Geo Trans DINOv2 74.3 81.0 87.6 19.7 1.9 50.8 67.4 76.0 41.8 4.9 63.7 73.9 76.2 86.2 0.4 Generative Geo Trans SD 77.2 84.0 89.9 16.5 1.8 51.3 68.7 78.4 35.6 4.8 65.2 76.1 78.7 71.0 0.4 Improvement 5.7 6.0 6.5 9.7 0.2 2.9 3.5 3.8 16.3 0.4 3.2 3.6 3.7 26.3 0.1 Match-Control Net Generative Predator Predator Overlap: 17.5% Inlier Ratio: 0.5% Overlap: 17.5% Inlier Ratio: 26.7% Figure 5. Left: The visualization of the generated RGB image pairs and the formed color source and target point clouds; Right: In low-overlap cases, the original Predator struggles with registration. By contrast, the Generative Predator, enhanced with generated color information, successfully align them well. Net training set (Dai et al., 2017) for model fine-tuning. Following the default fine-tuning configuration of Control Net (Zhang et al., 2023), we adopt the Adam W optimizer (Loshchilov, 2017) with a learning rate of 1e-5 and set the training epoch to 10. The code for this project is implemented in Py Torch, and all experiments are conducted on a server equipped with an Intel i5 2.2 GHz CPU and a TITAN RTX GPU. In our experiments, we integrate our zero-shot geometric-color fusion (Sec. 3.5) with three prevalent deep geometric descriptors: FCGF (Choy et al., 2019), Predator (Huang et al., 2021), and Geo Transformer (Qin et al., 2022), resulting in corresponding color-enhanced variants: Generative FCGF, Generative Predator, and Generative Geo Trans for method evaluation. Additionally, to validate our XYZ-RGB fusion scheme, we replace the real color point clouds (with actual RGB values) used by Color PCR (Mu et al., 2024) with our generated color point clouds (with synthesized RGB values), forming Generative Color PCR for 3D matching. Evaluation Metric. Following (El Banani et al., 2021; Yuan et al., 2023), we use rotation error, translation error, and Chamfer error, including the accuracy across varying thresholds and mean/median errors, for performance evaluation. 4.2. Comparison with Existing Methods Evaluation on Scan Net. We first perform model evaluation on a widely-used, large-scale indoor benchmark dataset, Scan Net (Dai et al., 2017). We follow the official data split to divide this dataset into the training, validation, and testing subsets, and construct view pairs by sampling image pairs that are 50 frames apart. Compared to the 20-frame separation used in (El Banani et al., 2021; Yuan et al., 2023), our approach with a 50-frame separation further reduces the overlap ratio (i.e., lower overlap), thereby increasing the registration difficulty. We compare our method against with one representative traditional descriptor: FPFH (Rusu et al., 2009), one scene-level end-to-end registration network: Reg TR (Yew & Lee, 2022), and five deep descriptors: FCGF (Choy et al., 2019), Predator (Huang et al., 2021), Geo Trans (Qin et al., 2022), Lepard (Li & Harada, 2022), and Ro ITr (Yu et al., 2023). We adopt RANSAC-50k as the pose estimator for FPFH, Lepard and Ro ITr, and select Generative Point Cloud Registration Table 2. Comparison of the methods on rotation, translation, and Chamfer distance on 3DMatch (Zeng et al., 2017) benchmark dataset. Rotation (deg) Translation (cm) Chamfer (mm) Accuracy Error Accuracy Error Accuracy Error Methods 5 10 45 Mean Med. 5 10 25 Mean Med. 1 5 10 Mean Med. FPFH (Rusu et al., 2009) 69.1 82.9 91.2 15.0 3.1 25.8 53.9 75.1 37.4 9.1 52.5 74.2 79.2 57.6 0.9 Lepard (Li & Harada, 2022) 84.3 91.0 94.1 11.1 2.1 43.1 75.2 88.9 21.8 5.8 72.1 88.3 90.5 45.3 0.4 Reg TR (Yew & Lee, 2022) 86.2 92.1 97.2 5.7 1.6 55.0 77.6 88.9 18.8 4.6 75.4 88.2 91.3 40.0 0.3 Ro ITr (Yu et al., 2023) 86.3 91.1 93.8 11.1 1.6 51.2 77.4 89.1 20.5 4.9 75.2 88.5 90.6 50.1 0.4 FCGF (Choy et al., 2019) 90.4 93.7 94.8 9.4 1.4 53.4 79.3 91.0 19.2 4.7 76.7 90.8 92.4 40.3 0.4 Generative FCGFDINOv2 91.5 94.3 95.3 8.5 1.4 53.6 79.3 91.5 18.1 4.6 77.5 91.1 92.7 41.1 0.4 Generative FCGFSD 94.3 96.7 98.1 4.5 1.4 54.3 81.5 93.1 12.5 4.7 78.2 92.9 94.6 37.7 0.4 Improvement 3.9 3.0 3.3 4.9 0.0 0.9 2.2 2.1 6.7 0.0 1.5 2.1 2.2 2.6 0.0 Predator (Huang et al., 2021) 85.0 91.5 94.2 10.5 2.0 42.1 72.5 87.1 22.6 5.8 71.2 85.8 88.6 45.0 0.5 Generative Predator DINOv2 88.1 94.8 96.9 6.2 1.8 44.7 73.9 88.4 15.5 5.6 72.4 87.7 90.8 33.1 0.4 Generative Predator SD 88.6 94.6 97.0 5.9 1.9 45.7 74.5 89.1 15.7 5.5 73.3 88.3 90.9 40.4 0.4 Improvement 3.6 3.1 2.8 4.6 0.2 3.6 2.0 2.0 7.1 0.3 2.1 2.5 2.3 11.9 0.1 Geo Trans (Qin et al., 2022) 88.9 91.8 93.3 12.0 1.4 59.8 81.0 90.1 24.6 4.0 79.2 89.0 90.6 53.3 0.3 Generative Geo Trans DINOv2 90.2 93.2 95.2 8.9 1.4 61.0 83.1 90.4 16.9 3.9 80.4 89.7 91.7 36.9 0.3 Generative Geo Trans SD 91.5 94.3 96.2 7.6 1.4 61.3 82.9 90.9 17.2 3.9 81.5 90.1 92.3 37.3 0.3 Improvement 2.6 2.5 2.9 4.4 0.0 1.5 2.1 0.8 7.7 0.1 2.3 1.1 1.7 16.4 0.0 SC2PCR (Chen et al., 2022), RANSAC-50k and LGR for (Generative) FCGF, (Generative) Predator, and (Generative) Geo Trans (Qin et al., 2022), to validate the robustness of our generative 3D registration paradigm across different pose estimators. Table 1 demonstrates that enhanced by the free-lunch color information generated by our Match Control Net, all generative versions of FCGF, Predator, and Geo Trans achieve significant performance improvements, such as 6.9% of Generative FCGF on 45 @Rotation metric. These confirm the generality and effectiveness of our proposed generative point cloud registration paradigm. Additionally, we find that compared to the DINOv2 image encoding, Stable Diffusion can capture more discriminative representations and achieve higher precisions. Evaluation on 3DMatch. We next evaluate our method on 3DMatch (Zeng et al., 2017), another widely-used benchmark dataset for 3D registration. We follow (El Banani et al., 2021; Yuan et al., 2023) as in Scan Net to produce the pairwise samples. Also, we increase the view separation from 20 to 40, resulting in point cloud pairs with lower overlap to increase the registration challenge. Table. 2 demonstrates that by incorporating FCGF, Predator, and Geo Trans into our generative point cloud registration framework, their generative variants also consistently achieve the performance gain, validating the effectiveness of our proposed paradigm. 4.3. Ablation Studies and Analysis Effectiveness of Match-Control Net. We first evaluate the performance contribution of our developed Match Control Net: (i) The top block of Table 3 demonstrates that, compared to using generated image pairs with only 2D3D geometric consistency (geo), incorporating both 2D-3D geometric consistency and cross-view texture consistency (geo+tex) through our Match-Control Net results in higher registration accuracy. This improvement is due to the additional benefit of consistent textures and colors, which further facilitate accurate correspondence identification. Additionally, we observe that the generated images with only 2D-3D geometric consistency can also bring performance gain in some criteria. We attribute it to that DINOv2 and Stable Diffusion can extract powerful semantic representations, mitigating the feature inconsistency of correspondences caused by the texture difference and thereby aiding correspondence identification. Furthermore, we visualize the generated image pair for given source and target point clouds in Fig. 5 (left). It shows that our Match-Control Net is capable of producing high-quality image pairs with consistent 2D-3D geometry and cross-view texture. Zero-Shot vs Finetuning. We further investigate the performance of Match-Control Net in both zero-shot and finetuned settings. As shown in the second block of Table 3, both approaches yield substantial improvements over FCGF. Moreover, because the finetuned Match-Control Net benefits from task-specific training, it consistently achieves higher registration accuracy than the zero-shot version. Notably, even few-shot finetuning with as few as 1K samples yields clear performance gains. Increasing the number of finetuning samples (e.g., to 3K or 5K) provides additional improvements; however, models trained on 3K or 5K samples show comparable registration accuracy in practice. Hence, we adopt 3K samples as our default finetuning configuration. Zero-Shot Geometric-Color Feature Fusion. We next conduct ablation studies on the zero-shot geometric-color feature fusion described in Eq. 6. As shown in the fourth block of Table 3, Generative Geo Trans exhibits varying Generative Point Cloud Registration Table 3. Ablation studies on 3DMatch (Zeng et al., 2017) dataset. (*) denotes the default configuration. Rotation (deg) Translation (cm) Chamfer (mm) Accuracy Error Accuracy Error Accuracy Error Methods 5 10 45 Mean Med. 5 10 25 Mean Med. 1 5 10 Mean Med. FCGF 90.4 93.7 94.8 9.4 1.4 53.4 79.3 91.0 19.2 4.7 76.7 90.8 92.4 40.3 0.4 Generative FCGFSD (geo) 92.4 96.1 97.8 5.2 1.5 53.3 79.2 91.7 13.5 4.8 75.6 91.1 93.1 35.8 0.4 Generative FCGFSD (geo + tex) 94.3 96.7 98.1 4.5 1.4 54.3 81.5 93.1 12.5 4.7 78.2 92.9 94.6 37.7 0.4 Generative FCGFSD (zero-shot) 92.4 96.1 97.3 5.4 1.5 54.3 80.3 92.4 13.0 4.6 77.3 92.1 94.0 33.7 0.4 Generative FCGFSD (finetuning) 94.3 96.7 98.1 4.5 1.4 54.3 81.5 93.1 12.5 4.7 78.2 92.9 94.6 37.7 0.4 Finetune (#samples=1000) 93.5 96.8 98.1 4.6 1.4 54.2 80.5 92.5 12.4 4.7 78.2 92.5 94.7 32.8 0.4 Finetune (#samples=3000)* 94.3 96.7 98.1 4.5 1.4 54.3 81.5 93.1 12.5 4.7 78.2 92.9 94.6 37.7 0.4 Finetune (#samples=5000) 93.6 97.2 98.0 4.4 1.5 54.0 80.8 92.7 11.9 4.5 77.7 92.9 94.2 32.3 0.4 Color PCR (Mu et al., 2024) 79.9 84.6 88.9 16.5 1.8 48.3 69.6 82.2 41.8 5.2 66.6 80.6 83.3 81.1 0.5 Generative Color PCR 83.6 89.8 93.2 12.0 1.9 47.3 73.3 86.7 28.3 5.3 70.3 85.7 88.2 59.1 0.4 Improvement 3.7 5.2 4.3 4.5 0.1 1.0 3.7 4.5 13.5 0.1 3.7 5.1 4.9 22.0 0.1 Color feat. dim. drgb = 16 90.8 93.9 96.3 7.6 1.4 62.2 83.4 90.7 18.1 3.9 81.5 89.7 91.7 41.8 0.3 Color feat. dim. drgb = 32 91.5 94.3 96.2 7.6 1.4 61.3 82.9 90.9 17.2 3.9 81.5 90.1 92.3 37.3 0.3 Color feat. dim. drgb = 64* 91.5 94.3 96.2 7.6 1.4 61.3 82.9 90.9 17.2 3.9 81.5 90.1 92.3 37.3 0.3 Color feat. dim. drgb = 128 91.0 94.3 96.0 8.1 1.4 62.0 83.4 91.6 18.1 3.8 81.5 90.4 92.4 41.2 0.3 Fusion weight ω = 0.0 88.3 93.3 96.6 6.9 1.6 53.7 77.0 86.5 20.8 4.7 74.6 85.9 88.5 50.5 0.4 Fusion weight ω = 0.25 90.7 95.5 97.2 6.1 1.5 57.9 81.5 89.8 16.5 4.2 78.6 89.2 92.0 40.4 0.3 Fusion weight ω = 0.50* 91.5 94.3 96.2 7.6 1.4 61.3 82.9 90.9 17.2 3.9 81.5 90.1 92.3 37.3 0.3 Fusion weight ω = 0.75 89.2 92.7 93.9 10.9 1.4 60.5 81.7 90.6 22.6 4.0 79.5 89.3 91.4 49.5 0.3 Fusion weight ω = 1.0 89.0 91.8 93.3 12.0 1.4 59.9 81.1 90.2 24.6 4.0 79.3 89.1 90.8 53.3 0.3 Real Color PC Generated Color PC Real RGB Image Generated RGB Image Mitigate Lighting Mitigate Calibration Errors Figure 6. Our Match-Control Net effectively mitigates calibration errors and lighting challenges commonly encountered in real-world RGB-D data, thereby improving the matching precision of color point cloud registration methods. registration performance under different color feature dimensions, drgb {16, 32, 64, 128}. We observe that a very small color feature dimension (e.g., drgb = 16) degrades performance due to limited semantic representational capacity, while excessively large dimensions do not yield significant performance gains. Therefore, to balance inference efficiency with registration precision, we set drgb = 64 as our default setting. Additionally, in the fifth block of Table 3, we investigate performance variations with different fusion weights ω {0.0, 0.25, 0.50, 0.75, 1.0}, where a larger ω places more emphasis on the geometric descriptors (see Eq. 6). Our results indicate that both overly high ω (which overemphasizes geometry) and overly low ω (which overemphasizes color) lead to degraded registration accuracy. By contrast, a balanced weight (e.g., ω = 0.50) achieves higher performance. As a result, we adopt ω = 0.50 as our default hyperparameter configuration. XYZ-RGB Fusion. We finally evaluate the effectiveness of the XYZ-RGB fusion (see Sec. 3.5). The third block in Table 3 demonstrates that, on the 3DMatch dataset, Genera- tive Color PCR with the synthesized color even outperforms the original Color PCR with the real color. This advantage is attributed not only to the high-quality pairwise image generation provided by our Match-Control Net, but also to several key benefits of our generated XYZ-RGB data over real XYZ-RGB data: (i) Mitigating calibration errors: As shown in Fig. 6 (left), some real RGB-D data would suffer from calibration errors, which may lead to misalignment in the colored point clouds. By contrast, our framework, benefiting from the powerful 2D-3D consistency generation ability, effectively reduces such calibration errors, producing higher-quality colored point clouds and enabling more accurate matching. (ii) Mitigating lighting challenges: Fig. 6 (right) shows that RGB data from real-world conditions can degrade under poor lighting, negatively impacting RGB-D matching performance. Our generative point cloud registration framework, however, can generate images with consistent lighting conditions, independent of real-world lighting issues, thus enhancing the overall lighting robustness. 5. Conclusion We have introduced a novel 3D registration paradigm, generative point cloud registration, which effectively leverages advanced 2D generative models to augment geometryonly 3D registration. To this end, we developed Match Control Net, a matching-specific variant of Control Net designed to synthesize paired RGB images for both source and target point clouds. By integrating depth-conditioned generation from Control Net, coupled conditional denoising, Generative Point Cloud Registration and coupled prompt guidance, these generated RGB image pairs preserve both 2D-3D geometric consistency and crossview texture consistency, thereby facilitating high-quality 3D matching. Notably, our generative framework is general and can be incorporated into a variety of registration methods to improve their performance. Extensive experiments demonstrate the effectiveness of the proposed paradigm. Impact Statement This paper presents work whose goal is to advance research in point cloud registration. While potential societal implications may exist, we do not find any that necessitate explicit discussion here. Acknowledgments This research is supported by the RIE2025 Industry Alignment Fund Industry Collaboration Projects (IAF-ICP) (Award I2301E0026), administered by A*STAR, as well as supported by Alibaba Group and NTU Singapore through Alibaba-NTU Global e-Sustainability Corp Lab (ANGEL). Ao, S., Hu, Q., Wang, H., Xu, K., and Guo, Y. Buffer: Balancing accuracy, efficiency, and generalizability in point cloud registration. In CVPR, 2023. Bae, K.-H. and Lichti, D. D. A method for automated registration of unorganised point clouds. Journal of Photogrammetry and Remote Sensing, 2008. Bai, X., Luo, Z., Zhou, L., Fu, H., Quan, L., and Tai, C.-L. D3feat: Joint learning of dense detection and description of 3d local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6359 6367, 2020. Besl, P. J. and Mc Kay, N. D. Method for registration of 3-D shapes. In Sensor fusion IV: control paradigms and data structures, volume 1611, pp. 586 606, 1992. Chen, S., Xu, H., Li, R., Liu, G., Fu, C.-W., and Liu, S. Sirapcr: Sim-to-real adaptation for 3d point cloud registration. In ICCV, 2023. Chen, Z., Sun, K., Yang, F., and Tao, W. Sc2-pcr: A second order spatial compatibility for efficient and robust point cloud registration. In CVPR, 2022. Chetverikov, D., Svirko, D., Stepanov, D., and Krsek, P. The trimmed iterative closest point algorithm. In Object recognition supported by user interaction for service robots, 2002. Choy, C., Park, J., and Koltun, V. Fully convolutional geometric features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8958 8966, 2019. Choy, C., Dong, W., and Koltun, V. Deep global registration. In CVPR, 2020. Dai, A., Chang, A. X., Savva, M., Halber, M., Funkhouser, T., and Nießner, M. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In CVPR, 2017. Deng, H., Birdal, T., and Ilic, S. Ppfnet: Global context aware local features for robust 3d point matching. In CVPR, 2018. El Banani, M., Gao, L., and Johnson, J. Unsupervisedr&r: Unsupervised point cloud registration via differentiable rendering. In CVPR, 2021. Fitzgibbon, A. W. Robust registration of 2D and 3D point sets. Image and vision computing, 2003. Fu, K., Liu, S., Luo, X., and Wang, M. Robust point cloud registration framework based on deep graph matching. In CVPR, 2021. Ge, X. Automatic markerless registration of point clouds with semantic-keypoint-based 4-points congruent sets. ISPRS Journal of Photogrammetry and Remote Sensing, 2017. Gressin, A., Mallet, C., Demantk e, J., and David, N. Towards 3D lidar point cloud registration improvement using optimal neighborhood knowledge. ISPRS journal of photogrammetry and remote sensing, 2013. Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Neur IPS, 2020. Huang, J., Kwok, T.-H., and Zhou, C. V4PCS: Volumetric 4PCS algorithm for global registration. Journal of Mechanical Design, 2017. Huang, S., Gojcic, Z., Usvyatsov, M., Wieser, A., and Schindler, K. Predator: Registration of 3d point clouds with low overlap. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4267 4276, 2021. Jiang, H., Shen, Y., Xie, J., Li, J., Qian, J., and Yang, J. Sampling network guided cross-entropy method for unsupervised point cloud registration. In ICCV, 2021a. Jiang, H., Xie, J., Qian, J., and Yang, J. Planning with learned dynamic model for unsupervised point cloud registration. ar Xiv preprint ar Xiv:2108.02613, 2021b. Generative Point Cloud Registration Jiang, H., Xie, J., and Yang, J. Action candidate based clipped double q-learning for discrete and continuous action tasks. In AAAI, 2021c. Jiang, H., Li, G., Xie, J., and Yang, J. Action candidate driven clipped double q-learning for discrete and continuous action tasks. IEEE Transactions on Neural Networks and Learning Systems, 2022. Jiang, H., Dang, Z., Wei, Z., Xie, J., Yang, J., and Salzmann, M. Robust outlier rejection for 3d registration with variational bayes. In CVPR, 2023a. Jiang, H., Salzmann, M., Dang, Z., Xie, J., and Yang, J. Se (3) diffusion model-based point cloud registration for robust 6d object pose estimation. Neur IPS, 2023b. Johnson, A. E. and Hebert, M. Using spin images for efficient object recognition in cluttered 3d scenes. IEEE Transactions on pattern analysis and machine intelligence, 1999. Langley, P. Crafting papers on machine learning. In Langley, P. (ed.), Proceedings of the 17th International Conference on Machine Learning (ICML 2000), pp. 1207 1216, Stanford, CA, 2000. Morgan Kaufmann. Li, J., Zhang, C., Xu, Z., Zhou, H., and Zhang, C. Iterative distance-aware similarity matrix convolution with mutual-supervised point elimination for efficient point cloud registration. In ECCV, 2020. Li, M., Yang, T., Kuang, H., Wu, J., Wang, Z., Xiao, X., and Chen, C. Controlnet++: Improving conditional controls with efficient consistency feedback. In ECCV, 2025. Li, Y. and Harada, T. Lepard: Learning partial point cloud matching in rigid and deformable scenes. In CVPR, 2022. Liu, S., Zhou, Y., Song, J., Zheng, T., Chen, K., Zhu, T., Feng, Z., and Song, M. Contrastive identity-aware learning for multi-agent value decomposition. In AAAI, 2023. Liu, S., Song, J., Zhou, Y., Yu, N., Chen, K., Feng, Z., and Song, M. Interaction pattern disentangling for multi-agent reinforcement learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024. Loshchilov, I. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. Mohamad, M., Rappaport, D., and Greenspan, M. Generalized 4-points congruent sets for 3D registration. In 3DV, 2014. Mohamad, M., Ahmed, M. T., Rappaport, D., and Greenspan, M. Super generalized 4pcs for 3D registration. In 3DV, 2015. Mu, J., Bie, L., Du, S., and Gao, Y. Colorpcr: Color point cloud registration with multi-stage geometric-color fusion. In CVPR, 2024. Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., Fernandez, P., Haziza, D., Massa, F., El Nouby, A., et al. Dinov2: Learning robust visual features without supervision. ar Xiv preprint ar Xiv:2304.07193, 2023. Qin, Z., Yu, H., Wang, C., Guo, Y., Peng, Y., and Xu, K. Geometric transformer for fast and robust point cloud registration. In CVPR, 2022. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In CVPR, 2022. Rusu, R. B., Blodow, N., Marton, Z. C., and Beetz, M. Aligning point cloud views using persistent feature histograms. In IROS, 2008. Rusu, R. B., Blodow, N., and Beetz, M. Fast point feature histograms (fpfh) for 3d registration. In ICRA, 2009. Salti, S., Tombari, F., and Di Stefano, L. Shot: Unique signatures of histograms for surface and texture description. Computer Vision and Image Understanding, 2014. Sharp, G. C., Lee, S. W., and Wehe, D. K. ICP registration using invariant features. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2002. Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020. Tombari, F., Salti, S., and Di Stefano, L. Unique shape context for 3d data description. In Proceedings of the ACM workshop on 3D object retrieval, 2010. Wang, H., Liu, Y., Wang, B., Sun, Y., Dong, Z., Wang, W., and Yang, B. Freereg: Image-to-point cloud registration leveraging pretrained diffusion models and monocular depth estimators. In ICLR, 2024. Wang, W., Mei, G., Ren, B., Huang, X., Poiesi, F., Van Gool, L., Sebe, N., and Lepri, B. Zero-shot point cloud registration. ar Xiv preprint ar Xiv:2312.03032, 2023. Wang, Y. and Solomon, J. M. Deep closest point: Learning representations for point cloud registration. In ICCV, 2019. Xu, Y., Boerner, R., Yao, W., Hoegner, L., and Stilla, U. Pairwise coarse registration of point clouds in urban scenes using voxel-based 4-planes congruent sets. ISPRS journal of photogrammetry and remote sensing, 2019. Generative Point Cloud Registration Yang, J., Li, H., and Jia, Y. Go-ICP: Solving 3D registration efficiently and globally optimally. In ICCV, 2013. Yang, L., Zhang, Z., Song, Y., Hong, S., Xu, R., Zhao, Y., Zhang, W., Cui, B., and Yang, M.-H. Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 2023. Yew, Z. J. and Lee, G. H. RPM-Net: Robust point matching using learned features. In CVPR, 2020. Yew, Z. J. and Lee, G. H. Regtr: End-to-end point cloud correspondences with transformers. In CVPR, 2022. Yu, H., Qin, Z., Hou, J., Saleh, M., Li, D., Busam, B., and Ilic, S. Rotation-invariant transformer for point cloud matching. In CVPR, 2023. Yuan, M., Fu, K., Li, Z., Meng, Y., and Wang, M. Pointmbf: A multi-scale bidirectional fusion network for unsupervised rgb-d point cloud registration. In ICCV, 2023. Zeng, A., Song, S., Nießner, M., Fisher, M., Xiao, J., and Funkhouser, T. 3dmatch: Learning local geometric descriptors from rgb-d reconstructions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1802 1811, 2017. Zhang, L., Rao, A., and Agrawala, M. Adding conditional control to text-to-image diffusion models. In ICCV, 2023. Generative Point Cloud Registration A. More Quantitative Analysis To further validate the effectiveness of our Generative Point Cloud Registration paradigm, we integrate the handcrafted geometric descriptor, FPFH, into our framework, forming Generative FPFH. As shown in Table 4, this generative variant achieves a significant performance improvement over the baseline FPFH on the 3DMatch benchmark dataset, regardless of whether DINOv2 encoding or Stable-Diffusion encoding is used. Table 4. Comparison of the methods on rotation, translation, and Chamfer distance on 3DMatch (Zeng et al., 2017) benchmark dataset. Rotation (deg) Translation (cm) Chamfer (mm) Accuracy Error Accuracy Error Accuracy Error Methods 5 10 45 Mean Med. 5 10 25 Mean Med. 1 5 10 Mean Med. FPFH (Rusu et al., 2009) 69.1 82.9 91.2 15.0 3.1 25.8 53.9 75.1 37.4 9.1 52.5 74.2 79.2 57.6 0.9 Generative FPFHDINOv2 83.7 91.4 95.9 7.9 2.1 33.9 64.0 80.5 24.7 6.9 62.7 80.6 83.2 50.0 0.6 Generative FPFHSD 88.2 94.7 96.9 6.1 2.0 37.6 68.9 85.5 18.5 6.6 66.9 85.7 88.8 41.2 0.6 Improvement 19.1 11.8 5.7 8.9 1.1 11.8 15.0 10.4 18.9 2.5 14.4 11.5 9.6 16.4 0.3 B. More Visualization Results of Match-Control Net In Fig.7, we present additional visualization results of the RGB image pairs generated by our Match-Control Net, along with the corresponding colorized source and target point clouds. Furthermore, Fig.8 visualizes the image pairs generated by our zero-shot Match-Control Net, while Fig. 9 showcases image pairs produced by the finetuned Match-Control Net. Match-Control Net Match-Control Net Match-Control Net Match-Control Net Figure 7. More visualization results of the generated RGB image pairs and the formed color source and target point clouds. Generative Point Cloud Registration Zero-Shot Match-Control Net Generation Figure 8. Source and target image generation via zero-shot Match-Control Net without any finetuning. Generative Point Cloud Registration Finetuned Match-Control Net Generation Figure 9. Source and target image generation via finetuned Match-Control Net.