# sagemix_saliencyguided_mixup_for_point_clouds__e00fb931.pdf Sage Mix: Saliency-Guided Mixup for Point Clouds Sanghyeok Lee Korea University cat0626@korea.ac.kr Minkyu Jeon Korea University jmk94810@korea.ac.kr Injae Kim Korea University dna9041@korea.ac.kr Yunyang Xiong Meta Reality Labs yunyang@fb.com Hyunwoo J. Kim Korea University hyunwoojkim@korea.ac.kr Data augmentation is key to improving the generalization ability of deep learning models. Mixup is a simple and widely-used data augmentation technique that has proven effective in alleviating the problems of overfitting and data scarcity. Also, recent studies of saliency-aware Mixup in the image domain show that preserving discriminative parts is beneficial to improving the generalization performance. However, these Mixup-based data augmentations are underexplored in 3D vision, especially in point clouds. In this paper, we propose Sage Mix, a saliency-guided Mixup for point clouds to preserve salient local structures. Specifically, we extract salient regions from two point clouds and smoothly combine them into one continuous shape. With a simple sequential sampling by re-weighted saliency scores, Sage Mix preserves the local structure of salient regions. Extensive experiments demonstrate that the proposed method consistently outperforms existing Mixup methods in various benchmark point cloud datasets. With Point Net++, our method achieves an accuracy gain of 2.6% and 4.0% over standard training in 3D Warehouse dataset (MN40) and Scan Object NN, respectively. In addition to generalization performance, Sage Mix improves robustness and uncertainty calibration. Moreover, when adopting our method to various tasks including part segmentation and standard 2D image classification, our method achieves competitive performance. Code is available at https://github.com/mlvlab/Sage Mix. 1 Introduction Deep neural networks have achieved high performance in various domains including image, video, and speech. Recent researches [1, 2, 3, 4, 5, 6] have been proposed to employ deep learning model for 3D vision, especially in point clouds. However, in the point cloud domain, deep learning models are prone to suffer from weak-generalization performance due to the limited availability of data compared to the image datasets, which contain almost millions of training samples. For alleviating the data scarcity issue, data augmentation is a prevalent solution to increase the training data. In the image domain, Mixup-based methods [7, 8, 9] combine two training images to generate augmented samples. More recent Mixup methods [10, 11, 12, 13] focus on leveraging saliency to preserve the discriminative regions such as foreground objects. Despite the success of saliencyaware Mixup methods, it has been less studied in point clouds due to its unordered and non-grid structure. Moreover, saliency-aware augmentation methods require solving an additional optimization problem [11, 12], resulting in a considerable computational burden. First two authors have equal contribution. Corresponding author. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Point Mixup Figure 1: Comparison of Mixup methods on point clouds. Given two original samples, (left) guitar and (right) chair, we generate samples by various Mixup methods. (Top) Point Mixup does not preserve the discriminative structure. (Middle) The samples generated by RSMix contain the local structure of each sample, but the discontinuity occurs at the border. (Bottom) Our method, Sage Mix, generates a continuous mixture preserving the local structure of original shapes. Recently, several works [14, 15, 16] have attempted to extend the concept of Mixup to point clouds. Point Mixup [14] enables the linear interpolation of point clouds based on optimal assignment. RSMix [15] proposed a shape-preserving Mixup framework that extracts and merges the rigid subsets of each sample. However, these approaches have limitations. Point Mixup generates samples without preserving the local structure of the original shapes. For example, in the top row of Figure 1, the structure of the guitar is not preserved in generated samples. RSmix generates discontinuous samples and these artifacts often hinder effective training. Furthermore, these methods disregard the saliency, thereby causing the loss of the discriminative local structure in the original point cloud. In this paper, we propose a Saliency-Guided Mixup for point clouds (Sage Mix) that preserves discriminative local structures and generates continuous samples with smoothly varying mixing ratios. For saliency estimation, we measure the contribution of each point by the gradients of a loss function. Through a simple sequential sampling via re-weighted saliency scores, Sage Mix samples the query points to extract salient regions without solving additional optimization problems. Then, Sage Mix smoothly combines point clouds, considering the distance to query points to minimize the loss of discriminative local structures as illustrated in Figure 1 (e.g., the neck of a guitar and the back of a chair, etc.). In summary, our contributions are fourfold: We propose a novel saliency-guided Mixup method for point clouds. To the best of our knowledge, this is the first work that utilizes saliency for Mixup in point clouds. We design a Mixup framework that preserves the salient local structure of original shapes while smoothly combining them into one continuous shape. Extensive experiments demonstrate that Sage Mix brings consistent and significant improvements over state-of-the-art Mixup methods in generalization performance, robustness, and uncertainty calibration. We demonstrate that the proposed method is extensible to various tasks including part segmentation and standard image classification. 2 Preliminaries In this section, we briefly review the basic concept of Mixup and summarize the variants of Mixup for images and point clouds (Table 1). Table 1: Variants of Mixup. Method Equation mixup [9] x = λxα + (1 λ)xβ Manifold mixup [7] h = λhα + (1 λ)hβ Cut Mix [8] x = M xα + (1 M) xβ Saliency Mix [13] x = M(xα) xα + (1 M(xα)) xβ Puzzle Mix [11] x = Z Π α xα + (1 Z) Π β xβ Co-Mixup [12] x = Pm i Zi xi Point cloud Point Mixup [14] P = {λpα i + (1 λ)pβ ϕ(i)}n i RSMix [15] P = (Pα Sα) Sβ α Sage Mix P = {λipα i + (1 λi)pβ ϕ(i)}n i Vicinal risk minimization with Mixup. Given observed data D = {(xi, yi)}m i and a function f : X Y, Chapelle et al. [17] learns the function f by minimizing the empirical vicinal risk: Rν(f) = 1 m Pm i=1 ℓ(f( xi), yi), where ℓis the loss function, and ( x, y) is the virtual featuretarget pair from the vicinal distribution ν of the observed data D. To construct an effective vicinal distribution in the image domain, Zhang et al. [9] introduced Mixup: x = λxα + (1 λ)xβ, y = λyα + (1 λ)yβ, (1) where (xα, yα), (xβ, yβ) are two pairs of data in training dataset D, and λ Beta(θ, θ) is a mixture ratio. Following [9], Verma et al. [7] proposed Manifold Mixup that applies Mixup in hidden representations (i.e., h = λhα + (1 λ)hβ) and Cut Mix [8] generates samples via cut-and-paste manner with binary mask M {0, 1}W H (i.e., x = M xα+(1 M) xβ, where indicates the element-wise product). Combining with saliency, Saliency Mix [13] improved Cut Mix by selecting the patch M(xα) with the maximum saliency values. Puzzle Mix optimizes the mask Z [0, 1]W H and transport Π for maximizing the saliency of the mixed sample (i.e., x = Z Π α xα+(1 Z) Π β xβ). Beyond two samples, Co-Mixup mixes multiple samples in a mini-batch by optimizing multiple masks (i.e., x = Pm i Zi xi). While saliency-aware Mixup methods boost the generalization of deep learning models, they are originally designed for the image domain. Hence, these methods are not directly applicable to point clouds due to the unordered and irregular structure. Mixup in point cloud. Several works [14, 15] tried to leverage the Mixup in point cloud. Point Mixup [14] linearly interpolates two point clouds by P = {λpα i + (1 λ)pβ ϕ (i)}n i , y = λyα + (1 λ)yβ, (2) ϕ = argmin ϕ Φ i pα i pβ ϕ(i) 2, (3) where Pt = {pt 1, ..., pt n} is the set of points with t {α, β}, n is the number of points, and ϕ : {1, ..., n} 7 {1, ..., n} is the optimal bijective assignment between two point clouds. In RSMix [15], they generate an augmented sample by merging the subsets of two objects, defined as P = (Pα Sα) Sβ α, where St Pt is the rigid subset and Sβ α denotes Sβ translated to the center of Sα. Although these methods have shown that Mixup is effective for point clouds, some limitations have remained unresolved: loss of original structures, discontinuity at the boundary, and loss of discriminative regions. Here, to address these issues, we propose a new Mixup framework in the following section. Figure 2: Illustration of Sage Mix pipeline. Given a pair of samples Pα, and Pβ, Sage Mix sequentially samples the query points qα, and qβ (green) based on saliency using Equation (4) and Equation (5), respectively. Then, using smoothly varying weights with respect to the distance to query points, Sage Mix generates an augmented sample P preserving the salient local structures. The main goal of Sage Mix is to generate an augmented sample that 1) maximally preserves the salient parts, 2) keeps the local structure of original shapes, and 3) maintains the continuity in the boundary. To achieve this goal, Sage Mix extracts salient regions centered at query points and smoothly combines them through continuous weights in Euclidean space. The overall pipeline is illustrated in Figure 2 and pseudocode is provided in Algorithm 1 . In this section, we delineate the process for selecting the query point based on the saliency map in Section 3.1 and Mixup methods for preserving the salient local structure in Section 3.2. 3.1 Saliency-guided sequential sampling We first introduce a query point qt Pt that is considered the center of the region for preserving the local shape. A naïve random sampling for a query point is simple, but it does not guarantee that it is placed in a salient region. Thus, we propose saliency-guided query point selection for maximally maintaining the salient part. We denote the saliency St of the input Pt by the norm of the gradient (i.e., St = Ptℓ(f(Pt), yt) ), following [11, 12]. Deterministically selecting a query point with a maximum saliency score (i.e., qt = pt i , where i = arg maxi(St i)) sounds promising but in practice, this makes it difficult to generate diverse samples since query points are always located at the same position. Moreover, if two query points qα, qβ are closely located in the Euclidean space, it is challenging to preserve the local structure of each sample because of the significant overlap. To address these issues, Sage Mix sequentially selects the query point based on the saliency scores. The first step is to extract the query point qα = pα i . To maximize the diversity, the query point is sampled with respect to the probability distribution defined as Pr(Iα = i) = sα i Pn i sα i , (4) where sα Sα is the saliency of pα i and Iα is a random variable for index (i Iα). That is, the points in a salient region have a high chance to be chosen as a query point. This simple sampling method efficiently provides diverse query points, thereby minimizing redundant selections. Given query point qα, the next step is to define a sampling method for qβ to alleviate the overlap between selected parts. We encourage the sampler to distance qβ from qα by reweighting the saliency scores: Pr(Iβ = i) = pβ i qα sβ i Pn i pβ i qα sβ i . (5) Algorithm 1 A saliency-guided Mixup for point clouds Input: Pα, Pβ, yα, yβ, Sα, Sβ, σ, ϕ, θ P = {p1, ..., pn} : set of points, y : target, S = {s1, ..., sn} : saliency values, σ : bandwidth, ϕ : assignment function, θ : shape parameter Output: P, y 1: Draw qα = pα i from Pα w.r.t. Pr(Iα = i) = sα i Pn i sα i Equation (4) 2: Draw qβ = pβ i from Pβ w.r.t. Pr(Iβ = i) = pβ i qα sβ i Pn i pβ i qα sβ i Equation (5) 3: for i = 1 to n do 4: wα i , wβ i Kσ(pα i , qα), Kσ(pβ i , qβ) Equation (6) 5: λi πwα i πwα i +(1 π)wβ ϕ(i) π Beta(θ, θ), Equation (9) 6: pi λipα i + (1 λi)pβ ϕ(i) Equation (8) 7: end for 8: λ 1 nΣn i λi 9: y λyα + (1 λ)yβ 10: return P = { p1, .., pn}, y Assuming the point pβ i qα, the probability of selecting the point pβ i decreases. This implies that Sage Mix samples a query point considering both distance and saliency. Finally, with Equation (4) and Equation (5), we can obtain the remotely located query points with high saliency scores, resulting in augmented samples preserving the discriminative structures of original shapes. 3.2 Shape-preserving continuous Mixup Since our objective is to preserve the region around the query point qt, we need to impose high weight on the points near the query point. Further, to alleviate the discontinuity between two samples, the weight should smoothly vary in Euclidean space. Herein, we use a Gaussian Radial Basis Function (RBF) kernel to calculate weights: wt i = Kσ(pt i, qt) = exp pt i qt 2 where wt i is the weight on the point pt i and σ R+ is a bandwidth for kernel. The weight wt i smoothly increases as the distance to qt decreases, resulting in the region around the query point being prone to respect its original shape. As shown in (Figure 2), the parts with higher weights maintain their original local structure more in a mixed sample, e.g., the body of the guitar and the tail of the airplane. Given two point clouds (Pα, Pβ) and their corresponding weights ({wα i }n i , {wβ i }n i ), we generate an augmented sample via point-wise interpolation. For differentially mixing points, we define the mixing ratio for the i-th point pair as λα i = wα i wα i + wβ ϕ(i) , λβ ϕ(i) = wβ ϕ(i) wα i + wβ ϕ(i) , (7) where ϕ is the assignment in Equation (3). Note, that the point-wise mixing ratio λα i is the ratio between the weight of two paired points (i.e., λβ ϕ(i) = 1 λα i ), thus enabling linear interpolation. For simplicity, we use the notation λi, (1 λi) instead of λα i , λβ ϕ(i). Then, we generate the virtual point cloud by modifying the Equation (2) to P = {λipα i + (1 λi)pβ ϕ(i)}n i , y = λyα + (1 λ)yβ, (8) where λ = 1 n Pn i λi can be interpreted as the overall mixing ratio that is used in label interpolation. In Mixup, the distribution of λ is a key factor for model training [9]. To control the distribution of λ, Table 2: 3D shape classification performance on MN40/OBJ_ONLY/PB_T50_RS. Model Method MN40 OBJ_ONLY PB_T50_RS Point Net [3] Base 89.2 79.1 65.4 + Point Mixup [14] 89.9 79.4 65.7 + RSMix [15] 88.7 77.8 65.7 + Sage Mix 90.3 79.5 66.1 Point Net++ [4] Base 90.7 86.5 79.7 + Point Mixup [14] 92.3 87.6 80.2 + RSMix [15] 91.6 87.4 81.1 + Sage Mix 93.3 88.7 83.7 Base 92.9 86.2 79.9 + Point Mixup [14] 92.9 86.9 82.5 + RSMix [15] 93.5 86.6 82.2 + Sage Mix 93.6 88.0 83.6 we introduce a prior factor π Beta(θ, θ): λi = πwα i πwα i + (1 π)wβ ϕ(i) . (9) Remarks. Our Sage Mix can emulate both Point Mixup and RSMix depending on the bandwidth σ. If the bandwidth σ is sufficiently large, Kσ(pt i, qt) Kσ(pt j, qt) = c, i =j, where c R+, resulting in λi πc πc+(1 π)c = π. That is, all points are mixed in the same ratio as Point Mixup. Conversely, when the bandwidth becomes smaller, Sage Mix changes the mixing ratio drastically around the boundary between the two shapes and generates an augmented sample like RSMix. Qualitative results are available in Section 4.2. See Section 4.2 for qualitative results. Additionally, with a minor change, Sage Mix is applicable in feature space as Manifold Mixup, see Appendix B.2 for more discussion. 4 Experiments In this section, we demonstrate the effectiveness of our proposed method Sage Mix with various benchmark datasets. First, for 3D shape classification, we evaluate the generalization performance, robustness, and calibration error in Section 4.1. Next, we provide an ablation study and analyses of Sage Mix in Section 4.2. Lastly, we study the extensibility of our method in Section 4.3. Implementation details are provided in Appendix A. Data. We use two benchmark dataset: 3D Warehouse dataset (MN40) [18] and Scan Object NN [19]. MN40 is a synthetic dataset containing 9,843 CAD models for training and 2,468 CAD models for evaluation. Each CAD model of MN40 is obtained from 3D Warehouse [20]. Scan Object NN, obtained from Scene NN [21] and Scan Net[22], is a real-world dataset that is split into 80% for training and 20% for evaluation. Among the variants of Scan Object NN, we adopt the simplest version (OBJ_ONLY) and the most challenging version (PB_T50_RS). For training models, we use only coordinates (x,y,z) of 1024 points without additional information such as the normal vector. Baselines. For a comparison with previous studies, we use three backbone models: Point Net [3], Point Net++ [4], and DGCNN [6]. We compare Sage Mix with the model under default augmentation in [3, 4, 6, 19] (Base), and other Mixup approaches (Point Mixup [14], RSMix [15]). We report Table 3: Robustness and calibration with DGCNN on OBJ_ONLY. Gaussian noise. Rotation 180 Scaling Dropout Calibration σ : 0.01 σ : 0.05 X-axis Z-axis 0.6 2.0 25% 50% ECE(%) DGCNN [6] 84.9 48.4 32.5 32.4 73.7 73.0 83.3 75.7 19.8 + Point Mixup [14] 85.0 61.3 31.7 32.7 73.8 73.0 84.2 74.9 6.8 + RSMix [15] 84.2 49.1 32.7 32.6 75.0 74.5 84.0 73.6 18.9 + Sage Mix 85.7 51.2 36.5 37.9 75.6 75.2 84.9 79.0 5.1 Table 4: Ablation of Saliency-guided sequential sampling. Metric DGCNN [6] Uniform Max Saliency only Sage Mix OA 86.2 86.8 86.1 87.8 88.0 the performance in overall accuracy. We highlight the best performance in red and second-best performance in yellow. 4.1 3D shape classification. Generalization performance. Table 2 summarizes the experimental results of 3D shape classification on three datasets. Our framework significantly outperforms all of the previous methods in every dataset and model. Although the datasets are quite saturated, the averages of the improvements against the Base with Point Net, Point Net++, and DGCNN are 0.7%, 2.9%, and 2.1%, respectively. With Point Net++, Sage Mix improves the overall accuracy by 2.6%, 2.2%, and 4.0% compared to Base in MN40, OBJ_ONLY, and PB_T50_RS, respectively. We observe similar performance improvements in DGCNN by 0.7%, 1.8%, and 3.7% over Base. These consistent improvements demonstrate the effectiveness of our framework. Robustness. We adopt DGCNN and OBJ_ONLY to evaluate the robustness of models trained by our method. We compare our method with previous methods [14, 15] on four types of corruption: (1) jittering the point cloud with Gaussian noise (σ 0.01, 0.05), (2) Rotation 180 , (3) Scaling with a factor in {0.6, 2.0}, and (4) Dropout 25% or 50% of all points. As shown in Table 3, Sage Mix consistently improves the robustness in various corruption. Except for Gaussian noise with σ = 0.05, DGCNN trained with Sage Mix shows the best robustness with significant gains compared to previous methods. Specifically, Sage Mix outperforms the other methods with gains of 5.5% for rotation (Z-axis), 2.2% for scaling ( 2.0), and 3.3% for dropout (50%) over base DGCNN. Calibration. Previous works [23, 24] have proven that deep neural networks tend to be overconfident resulting in poorly calibrated models. In other words, a well-calibrated model should provide an accurate probability according to its predictions. Here, we report Expected Calibration Error (ECE) [23] to estimate the uncertainty calibration of models. We use the same setting as the robustness test for reporting ECE (Table 3). Overall, this result reveals that Sage Mix provides the best performance on uncertainty calibration compared to the other methods. See Appendix B.3 for more results. 4.2 Ablation study and analyses We provide various quantitative and qualitative analyses for a better understanding of Sage Mix. We use DGCNN and OBJ_ONLY for an ablation study and MN40 for visualization. Ablation on query point sampling. We explore the effectiveness of saliency-guided sequential query point sampling. Table 4 shows the results with various sampling methods for query points. A naïve uniform sampling for query point (Uniform), without considering the saliency, introduces +0.6% gains over base DGCNN. Interestingly, when Sage Mix (deterministically) selects the query point with maximum saliency values (Max), the performance is even degraded (-0.1%). With our Figure 3: Qualitative analyses on prior factor π and the bandwidth σ. Given two samples (left) table and airplane, Sage Mix generates a sample based on (Top) various π with a fixed σ = 0.3 and (Bottom) various σ with a fixed π = 0.5. Note that the green points indicate query points. Table 5: Quantitative analysis on the bandwidth σ σ 0.1 0.3 0.5 1.0 2.0 OA 87.2 88.0 87.6 87.3 87.6 saliency-guided random sampling for two query points based only on Equation (4) (Saliency Only), the performance is significantly improved to 87.8% (+1.6%). These results suggest that although saliency is key to effective training, we also have to consider the diversity for generating a new sample. Finally, the best accuracy of 88.0% (+1.8%) is obtained by our saliency-guided sequential sampling considering saliency and the distance between query points together. Prior factor. We introduce the parameter π as a prior factor for mixing ratio λ. The top row of Figure 3 is the visualization of samples generated by Sage Mix with various prior factor π. Given two samples and their corresponding query points colored in green, Sage Mix generates a sample with the salient local regions from both samples, e.g., the head and right-wing of an airplane and the back of a table are preserved. We observe that the generated sample gets close to the airplane as π 1, and vice versa. In short, Sage Mix controls the distribution of mixing ratio λ based on π. Bandwidth. The bandwidth σ of the RBF kernel controls the change of point-wise mixing ratios in Sage Mix. As mentioned in Section 4.2, when the bandwidth is sufficiently large (σ = 2), Sage Mix emulates Point Mixup [14]. Sage Mix tends to yield globally even weights and constant mixing ratios for all points rather than focusing on local parts. In contrast, as the bandwidth gets smaller, Sage Mix tends to impose higher weights around the query point preserving the local structure more precisely. These are well exemplified in the bottom row of Figure 3. For instance, when σ = 0.25, we notice the steep change in the boundary of two samples while preserving the salient local structure. This allows generating augmented samples like RSMix. In short, our Sage Mix can exhibit similar behaviors as Point Mixup and RSMix depending on the bandwidth σ. We also share the quantitative analysis of the bandwidth with DGCNN and OBJ_ONLY in Table 5. We observed that Sage Mix with a wide range of bandwidth (0.1 to 2.0) consistently outperforms previous Mixup methods (e.g., 86.9%, 86.6% for Point Mixup, RSMix). 4.3 Extensions to part segmentation and 2D image classification. Part segmentation. For part segmentation, we train DGCNN on 3D Warehouse (SN-Parts) [20, 18]. In part segmentation, since a model predicts a label for each point pi, we generate point-wise ground truth, i.e., yi = λiyα i + (1 λi)yβ ϕ(i). Also, for a comparison with previous methods, we used the official code by the authors of Point Mixup and RSMix with minor modifications for generating Table 6: Part segmentation performance on SN-Parts [18]. Model Base Point Mixup [14] RSMix [15] Sage Mix Point Net++ [4] 85.1 85.5 85.4 85.7 DGCNN [6] 85.1 85.3 85.2 85.4 Table 7: 2D image classification performance with Pre Act Res Net18 [25] on CIFAR-100. Dataset Vanilla Mixup Manifold Cut Mix Saliency Mix Puzzle Mix Co-Mixup Sage Mix CIFAR-100 76.41 77.57 78.36 78.71 79.06 79.38 80.13 80.16 point-wise ground truth. We follow the settings in [4, 6] to evaluate our method and reports the results in Table 6. Note that although the gain seems small, Sage Mix outperforms previous Mixup methods. Also, considering the already saturated performance, we believe that the improvement (+0.6%, +0.3% in Point Net++, DGCNN) over the base model is noteworthy. 2D classification. Our framework is also applicable to 2D image classification. Following [12], we used Pre Act Res Net18 [25] in the CIFAR-100 for our experiments. We compare our method with several Mixup baselines in the image domain [11, 12, 13, 7, 9, 8]. Although our method is designed for point clouds, it shows competitive performance. Sage Mix achieves the best accuracy of 80.16% with Pre Act Res Net18 on the CIFAR-100 dataset (Table 7). It is worth noting that Puzzle Mix and Co-Mixup require an additional optimization which introduces a considerable computational overhead. Specifically, Sage Mix is 6.05 faster than Co-Mixup [12] per epoch in the CIFAR-100 dataset. We believe that our simple sampling technique is helpful to improve the generalization power of the model. For more details, see Appendix B.4. 5 Related works Deep learning on point clouds. Point Net [3] is a pioneering work that designs a novel deep neural network for processing unordered 3D point sets with a multi-layer perceptron. Inspired by CNNs, Qi et al. [4] propose Point Net++ with a hierarchical architecture. In DGCNN, Wang et al. [6] introduce Edge Conv which utilizes edge features from the dynamically updated graph. Additionally, various works have focused on point-wise multi-layer perceptron [26, 27, 28], convolution [29, 30, 31, 1, 2, 5, 32], and graph-based methods [33, 34] to process point clouds. Parallel to these approaches, other recent works [14, 35, 15, 36, 8] focus on data augmentation to improve the generalization power of deep neural networks in point clouds. Mixup. Mixup [9] is a widely used regularization technique, which linearly interpolates a pair of images to generate an augmented sample. Following this work, Verma et al. [7] propose Manifold Mixup that extends Mixup to the hidden representations. Cut Mix [8] replaces a part of an image with a part of another one. More recent studies [11, 12, 13] have been proposed to preserve the saliency while mixing samples. In point clouds, Point Mixup [14] is the first approach that adapts the concept of Mixup in point clouds with the optimal assignments. Instead of linear interpolation, Lee et al. [15] propose RSMix which merges the subsets of two point clouds inspired by Cut Mix. Saliency. Measuring the saliency of data using neural networks has been studied to obtain a more precise saliency map [37, 38, 39]. The saliency has been prevalent in various fields such as object detection, segmentation, and speech recognition [10, 40, 41, 42, 43]. Similarly, Point Cloud Saliency Map [44] constructed the saliency map to identify the critical points via building a gradient-based saliency map. Recently, saliency has been used in Mixup framework [11, 12, 13] to prevent generating samples only with background or irrelevant regions to the target objects. 6 Conclusion We propose Sage Mix, a novel saliency-guided Mixup for point clouds to preserve salient local structures. Our method generates an augmented sample with a continuous boundary while preserving the discriminative regions. Additionally, with a simple saliency-guided sequential sampling, Sage Mix achieves state-of-the-art performance in various metrics (e.g., generalization, robustness, and uncertainty calibration). Moreover, we demonstrate that the proposed method is extensible to various tasks: part segmentation and standard 2D image classification. The visualization supports that Sage Mix generates a continuous mixture while respecting the salient local structure. Acknowledgments and Disclosure of Funding This work was partly supported by ICT Creative Consilience program (IITP-2022-2020-0-01819) supervised by the IITP; the National Supercomputing Center with supercomputing resources including technical support (KSC-2022-CRE-0261); and IITP grant funded by the Korea government (MSIT) (No.2021-0-02312, Efficient Meta-learning Based Training Method and Multipurpose Multi-modal Artificial Neural Networks for Drone AI). [1] Yangyan Li, Rui Bu, Mingchao Sun, Wei Wu, Xinhan Di, and Baoquan Chen. Pointcnn: Convolution on x-transformed points. Neur IPS, 2018. [2] Yongcheng Liu, Bin Fan, Shiming Xiang, and Chunhong Pan. Relation-shape convolutional neural network for point cloud analysis. In CVPR, 2019. [3] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In CVPR, 2017. [4] Charles Ruizhongtai Qi, Li Yi, Hao Su, and Leonidas J Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Neur IPS, 2017. [5] Hugues Thomas, Charles R Qi, Jean-Emmanuel Deschaud, Beatriz Marcotegui, François Goulette, and Leonidas J Guibas. Kpconv: Flexible and deformable convolution for point clouds. In ICCV, 2019. [6] Yue Wang, Yongbin Sun, Ziwei Liu, Sanjay E Sarma, Michael M Bronstein, and Justin M Solomon. Dynamic graph cnn for learning on point clouds. TOG, 2019. [7] Vikas Verma, Alex Lamb, Christopher Beckham, Amir Najafi, Ioannis Mitliagkas, David Lopez Paz, and Yoshua Bengio. Manifold mixup: Better representations by interpolating hidden states. In ICML, 2019. [8] Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In ICCV, 2019. [9] Hongyi Zhang, Moustapha Cissé, Yann N. Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In ICLR, 2018. [10] Chengyue Gong, Dilin Wang, Meng Li, Vikas Chandra, and Qiang Liu. Keepaugment: A simple information-preserving data augmentation approach. In CVPR, 2021. [11] Jang-Hyun Kim, Wonho Choo, and Hyun Oh Song. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In ICML, 2020. [12] Jang-Hyun Kim, Wonho Choo, Hosan Jeong, and Hyun Oh Song. Co-mixup: Saliency guided joint mixup with supermodular diversity. In ICLR, 2021. [13] A. F. M. Shahab Uddin, Mst. Sirazam Monira, Wheemyung Shin, Tae Choong Chung, and Sung Ho Bae. Saliencymix: A saliency guided data augmentation strategy for better regularization. In ICLR, 2021. [14] Yunlu Chen, Vincent Tao Hu, Efstratios Gavves, Thomas Mensink, Pascal Mettes, Pengwan Yang, and Cees GM Snoek. Pointmixup: Augmentation for point clouds. In ECCV, 2020. [15] Dogyoon Lee, Jaeha Lee, Junhyeop Lee, Hyeongmin Lee, Minhyeok Lee, Sungmin Woo, and Sangyoun Lee. Regularization strategy for point cloud via rigidly mixed sample. In CVPR, 2021. [16] Jinlai Zhang, Lyujie Chen, Bo Ouyang, Binbin Liu, Jihong Zhu, Yujin Chen, Yanmei Meng, and Danfeng Wu. Pointcutmix: Regularization strategy for point cloud classification. Neurocomputing, 2022. [17] Olivier Chapelle, Jason Weston, Léon Bottou, and Vladimir Vapnik. Vicinal risk minimization. Neur IPS, 2000. [18] Chenfeng Xu, Shijia Yang, Bohan Zhai, Bichen Wu, Xiangyu Yue, Wei Zhan, Peter Vajda, Kurt Keutzer, and Masayoshi Tomizuka. Image2point: 3d point-cloud understanding with pretrained 2d convnets. ar Xiv preprint ar Xiv:2106.04180, 2021. [19] Mikaela Angelina Uy, Quang-Hieu Pham, Binh-Son Hua, Thanh Nguyen, and Sai-Kit Yeung. Revisiting point cloud classification: A new benchmark dataset and classification model on real-world data. In ICCV, 2019. [20] Sketch Up. 3D Warehouse, 2022. URL https://3dwarehouse.sketchup.com/. [21] Binh-Son Hua, Quang-Hieu Pham, Duc Thanh Nguyen, Minh-Khoi Tran, Lap-Fai Yu, and Sai-Kit Yeung. Scenenn: A scene meshes dataset with annotations. In 3DV, 2016. [22] Angela Dai, Angel X Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Nießner. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In CVPR, 2017. [23] Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. In ICML, 2017. [24] Aviral Kumar, Sunita Sarawagi, and Ujjwal Jain. Trainable calibration measures for neural networks from kernel mean embeddings. In ICML, 2018. [25] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Identity mappings in deep residual networks. In ECCV, 2016. [26] Mor Joseph-Rivlin, Alon Zvirin, and Ron Kimmel. Momen (e) t: Flavor the moments in learning to classify shapes. In ICCV, 2019. [27] Xu Yan, Chaoda Zheng, Zhen Li, Sheng Wang, and Shuguang Cui. Pointasnl: Robust point clouds processing using nonlocal neural networks with adaptive sampling. In CVPR, 2020. [28] Hengshuang Zhao, Li Jiang, Chi-Wing Fu, and Jiaya Jia. Pointweb: Enhancing local neighborhood features for point cloud processing. In CVPR, 2019. [29] Binh-Son Hua, Minh-Khoi Tran, and Sai-Kit Yeung. Pointwise convolutional neural networks. In CVPR, 2018. [30] Shiyi Lan, Ruichi Yu, Gang Yu, and Larry S Davis. Modeling local geometric structure of 3d point clouds using geo-cnn. In CVPR, 2019. [31] Huan Lei, Naveed Akhtar, and Ajmal Mian. Octree guided cnn with spherical kernels for 3d point clouds. In CVPR, 2019. [32] Wenxuan Wu, Zhongang Qi, and Li Fuxin. Pointconv: Deep convolutional networks on 3d point clouds. In CVPR, 2019. [33] Yiru Shen, Chen Feng, Yaoqing Yang, and Dong Tian. Mining point cloud local structures by kernel correlation and graph pooling. In CVPR, 2018. [34] Yingxue Zhang and Michael Rabbat. A graph-cnn for 3d point cloud classification. In ICASSP, 2018. [35] Sihyeon Kim, Sanghyeok Lee, Dasol Hwang, Jaewon Lee, Seong Jae Hwang, and Hyunwoo J Kim. Point cloud augmentation with weighted local transformations. In ICCV, 2021. [36] Ruihui Li, Xianzhi Li, Pheng-Ann Heng, and Chi-Wing Fu. Pointaugment: an autoaugmentation framework for point cloud classification. In CVPR, 2020. [37] Ramprasaath R Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, and Dhruv Batra. Grad-cam: Visual explanations from deep networks via gradient-based localization. In ICCV, 2017. [38] Lijun Wang, Huchuan Lu, Xiang Ruan, and Ming-Hsuan Yang. Deep networks for saliency detection via local estimation and global search. In CVPR, 2015. [39] Rui Zhao, Wanli Ouyang, Hongsheng Li, and Xiaogang Wang. Saliency detection by multicontext deep learning. In CVPR, 2015. [40] Chanho Jung and Changick Kim. A unified spectral-domain approach for saliency detection and its application to automatic object segmentation. TIP, 2011. [41] Ozlem Kalinli and Shrikanth S Narayanan. A saliency-based auditory attention model with applications to unsupervised prominent syllable detection in speech. In INTERSPEECH, 2007. [42] Zhixiang Ren, Shenghua Gao, Liang-Tien Chia, and Ivor Wai-Hung Tsang. Region-based saliency detection and its application in object recognition. TCSVT, 2013. [43] Yunchao Wei, Jiashi Feng, Xiaodan Liang, Ming-Ming Cheng, Yao Zhao, and Shuicheng Yan. Object region mining with adversarial erasing: A simple classification to semantic segmentation approach. In CVPR, 2017. [44] Tianhang Zheng, Changyou Chen, Junsong Yuan, Bo Li, and Kui Ren. Pointcloud saliency maps. In ICCV, 2019. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] See abstract and Section 1. (b) Did you describe the limitations of your work? [Yes] See supplement. (c) Did you discuss any potential negative societal impacts of your work? [Yes] See supplement. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See supplement. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See supplement. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See supplement. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See supplement. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] See supplement. (b) Did you mention the license of the assets? [Yes] See supplement. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] See supplement. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] We use the publicly available benchmark datasets. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] We use the publicly available benchmark datasets. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]