# domain_generalization_with_vital_phase_augmentation__fd8ecb65.pdf Domain Generalization with Vital Phase Augmentation Ingyun Lee, Wooju Lee, Hyun Myung* Urban Robotics Lab, School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Republic of Korea {iglee, dnwn24, hmyung}@kaist.ac.kr Deep neural networks have shown remarkable performance in image classification. However, their performance significantly deteriorates with corrupted input data. Domain generalization methods have been proposed to train robust models against out-of-distribution data. Data augmentation in the frequency domain is one of such approaches that enable a model to learn phase features to establish domain-invariant representations. This approach changes the amplitudes of the input data while preserving the phases. However, using fixed phases leads to susceptibility to phase fluctuations because amplitudes and phase fluctuations commonly occur in outof-distribution. In this study, to address this problem, we introduce an approach using finite variation of the phases of input data rather than maintaining fixed phases. Based on the assumption that the degree of domain-invariant features varies for each phase, we propose a method to distinguish phases based on this degree. In addition, we propose a method called vital phase augmentation (VIPAug) that applies the variation to the phases differently according to the degree of domain-invariant features of given phases. The model depends more on the vital phases that contain more domain-invariant features for attaining robustness to amplitude and phase fluctuations. We present experimental evaluations of our proposed approach, which exhibited improved performance for both clean and corrupted data. VIPAug achieved SOTA performance on the benchmark CIFAR-10 and CIFAR-100 datasets, as well as near-SOTA performance on the Image Net-100 and Image Net datasets. Our code is available at https://github.com/excitedkid/vipaug. Introduction Deep learning is being actively explored for various applications in computer vision such as image classification and object detection (He et al. 2016; Tan and Le 2019; Dosovitskiy et al. 2021). The rapid development of deep learning methods has led to performance that can surpass that of human effort on some tasks. For example, deep neural networks (DNNs) can achieve high accuracy on image classification tasks with in-distribution data. However, the realworld performance of DNNs can be poor compared with that of manual classification by humans (Hendrycks and Dietterich 2019). Because the distributions of train and test *The corresponding author. Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Orignial Amplitude Phase Corrupted Amplitude Phase Reconstructed (a) Reconstructed (b) Figure 1. Amplitude and phase of original and blur images, with reconstructed images. Amplitude and phase fluctuations can be observed in II compared with I. III-(a) retains vital phases of the original image and sets other phases to 0. In contrast, III-(b) retains the non-vital phases at the same ratio and sets other phases to 0. In III, the amplitude of the original image is kept unchanged. The amplitude and nonvital phases are shown to contain less domain-invariant features. datasets may differ in the real world, deep learning models cannot be trained to compensate for all of the potential types of data corruption. To address this challenge, domain generalization methods have been developed to train models to be more robust to out-of-distribution (OOD) data (Lee and Myung 2022). These techniques aim to minimize any deterioration in performance on clean data while improving the performance of deep learning models on corrupted data. Data augmentation methods have also been proposed to improve domain generalization. Some of these approaches (Chen et al. 2021; Xu et al. 2021) based on the frequency domain show that phases contain domain-invariant features. To make the models depend on the phases, only the amplitudes of the input data are varied with several differ- The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) ent techniques and fix the phases. The fixed phases are then combined with the augmented amplitudes to reconstruct the image. However, with corrupted data, amplitudes and phases can fluctuate significantly as shown in Figure 1 and Figure 2. Therefore, existing methods with fixed phases are not robust to phase fluctuations. To address the limitations of existing methods, we propose to introduce finite phase variations to ensure robustness to phase fluctuations. We propose two hypotheses. First, the degree of domain-invariant feature inclusion, which we define as the robustness weight, varies for each phase. We define a phase with relatively high robustness weights as a vital phase and a phase with low robustness weights as a nonvital phase. Accordingly, we propose a method to detect vital and non-vital phases based on the magnitude of the amplitudes. Second, applying different strengths of variations according to robustness weights allows a model to depend more on vital phases, which enhances its robustness against corruption. By retaining the advantages of existing methods and addressing the vulnerability to phase fluctuations, we propose a novel approach called vital phase augmentation (VIPAug). VIPAug applies variations to the phases of input data based on robustness weight and replaces all amplitudes, enabling the model to depend on the vital phases. VIPAug incorporates phase variations by employing a Gaussian distribution and partially replacing the phases with those of fractal images. We also present the experimental results of our approach, which show the improved accuracy on both clean and corrupted data compared with baseline methods. The contributions of this study are summarized as follows: We experimentally demonstrate that the robustness weights of phases differ for the first time. We propose a method to identify vital and non-vital phases based on their weights. We propose VIPAug as a novel augmentation approach that combines the new phase variations with existing methods based on amplitudes variations. This approach enables the model to perform more robustly against phase fluctuations while depending on the phases. Our experimental results show that the proposed method achieved state-of-the-art performance on the CIFAR-10 and CIFAR-100 datasets (Krizhevsky and Hinton 2009) and nearly state-of-the-art performance on the Image Net100 and Image Net datasets (Deng et al. 2009). Related Work Domain Generalization Deep learning models should be robust against unseen domains that may be used in real-world applications. Domain generalization methods aim to generalize models to OOD data by using only training data from a given source domain. Domain generalization can be implemented in a variety of ways, including contrastive learning, ensemble learning, and meta-learning. Contrastive learning methods reduce the multi-domain gap to improve generalization ability. Motiian et al. (2017) exploited the Siamese architecture Figure 2. The number of fluctuated phases from clean to corrupted domains. We extracted an arbitrary image from Image Net-C and calculated the average value of all corruption types at corruption severity level 3. Phase fluctuations above the threshold value were counted. The range of phase is [ π, π]. Most phases were changed even when the threshold was small. with a contrastive loss. Yoon et al. (2019) extended a contrastive semantic alignment loss to mitigate the bias of data and establish domain-invariant representations. Ensemble learning methods combine several models to improve generalization. Ding et al. (2017) used multiple domain-specific deep neural networks to capture a shared representation within multiple sources. Similarly, Liu et al. (2020) proposed a multi-site network with domainspecific batch normalization layers. The meta-learning approach diversifies different models to improve their stability and generalization performance. Zhao et al. (2021) proposed a memory-based identification loss designed to harmonize with meta-learning. All these methods have a limitation in that they do not directly increase the diversity of the training data. For this reason, we concentrate on data augmentation among various methods for domain generalization. Data Augmentation Data augmentation has been studied to improve the generalization performance of deep learning models. Mixup (Zhang et al. 2018) is designed to mix two images with linear combinations to improve generalization ability. Cutout (De Vries and Taylor 2017) and Random Erasing (Zhong et al. 2020) randomly erases a part of an image to improve accuracy and generalize to the occluded objects. Auto Augment (Cubuk et al. 2019) optimizes a group of augmentations with reinforcement learning. However, these methods only generalize a model to limited scenarios and are not robust to various distributional shifts such as common corruptions. Common corruptions refer to the possible distortions and distributional shifts in the real world such as shot noise, motion blur, and snow. Recently, several data augmentation methods have been proposed to improve performance on common corruption scenarios. These methods generate multi-source domains from a single-source domain using various transformations and mixing strategies. Aug Mix (Hendrycks et al. 2019) proposed parallel data pipelines to generate diverse domains while maintaining semantic content. Pix Mix (Hendrycks et al. 2022) mixes original The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Phase variations Filter i DFT Partial Replacement Figure 3. Overall structure of VIPAug. VIPAug contains VIPAug-G and VIPAug-F. VIPAug-G introduces phase variations by using Gaussian distributions with different variances, σ2 vital and σ2 nonvital . VIPAug-F employs the phases of the fractal images. We introduce finite phase variations by finding vital phases with a filter and applying variations with different strengths depending on the robustness weight. A, P, and P denote the amplitude, phase, and varied phase spectrum. images with external fractal images to introduce greater structural complexity. Zhou et al. (2021) randomly mixed instance-level feature statistics of training samples across source domains. However, these methods do not take into account that the image phases contain domain-invariant features. Data augmentation using the frequency domain has also become a topic of active research that leverages domaininvariant features in images. APR-SP (Chen et al. 2021) fixes the phases and replaces the amplitudes with those from other images. FACT (Xu et al. 2021) fixes the phases and mixes the amplitudes with those from other images. Both methods introduce amplitude variations to enable a model to learn domain-invariant features from the phases. However, fixing the phases makes DNNs vulnerable to phase fluctuations. Hybrid Augment++ (Yucel, Cinbis, and Duygulu 2023) makes the model rely on the low-frequency components of data, but this approach does not consider variations in the robustness weight of each phase. PRIME (Modas et al. 2022) is an integrated method that considered augmentation in the spectral, spatial, and color domains. Although this approach greatly increased diversity with augmentation in the three domains, the authors did not consider that the phase contains domain-invariant features in an image. Method We propose VIPAug as a data augmentation method that integrates changes in the amplitudes and finite variations in the phase spectrum. The phase variations apply different intensities of variations to vital and non-vital phases according to their robustness weights. First, we propose a method to distinguish the vital and non-vital phases using the magnitude of the amplitudes. VIPAug contains two types of phase augmentation; one utilizes Gaussian distributions, and the other employs fractal images. VIPAug encourages the model to depend on the phases over the amplitudes, specifically on the vital phases. Due to this dependence on the vital phases, the model achieves robustness against fluctuations in terms of both amplitudes and phases. The entire VIPAug process is shown in Figure 3. Detection of Vital Phase The conventional approach uses 2D discrete Fourier transform (DFT) for each channel of an RGB image to obtain amplitudes and phases. Unlike 2D DFT, 3D DFT can be used to acquire amplitudes and phases that include features between each channel. Leveraging these amplitude and phase spectrums can improve the accuracy on clean and corrupted data. With image s height H, width W, channel C, coordinates of image s spatial domain (x, y, z) and frequency domain (u, v, w), the 3D DFT equation is represented as follows: F(u, v, w) = z=0 fe j2π( x where the input image is represented by f = f(x, y, z). We can derive the image s amplitudes A(u, v, w) and phases P(u, v, w): F(u, v, w) = |F(u, v, w)|ej arctan I(u,v,w) = A(u, v, w)ej P (u,v,w), (2) where I(u, v, w) and R(u, v, w) represent the imaginary and real parts of the DFT result. The relation between the image f and its corresponding amplitudes A(u, v, w) and phases P(u, v, w) is described using inverse discrete Fourier transform (i DFT): w=0 Aej{2π( u (3) where A = A(u, v, w) and P = P(u, v, w). The image f can be represented as a linear combination of complex exponential terms. The amplitude of the exponential term is The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Amplitude Spectrum argmax function Phase Spectrum u v v vital phase Figure 4. Vital phase detection method. The vital phase coordinates are found by applying the argmax function to the filter size of the amplitude spectrum region. The filter moves on the amplitude spectrums without overlaps. This figure is shown in the case of a gray scale image. proportional to the number of object features. Object features are semantically preserved across domains. Therefore, the phase containing more domain-invariant features has a larger amplitude. We hypothesize that the model should depend more on phases at larger amplitudes than relatively lower ones. The vital phase coordinates (uvital, vvital, wvital) are determined by applying an S S 1 argmax filter to the amplitudes of each region as shown in Figure 4. The filter encompasses all regions without any overlap. For sets Uvital , Vvital , and Wvital consisting of the elements uvital, vvital, and wvital, respectively, we get vital phase coordinate set Cvital, where Cvital = {(u, v, w) | u Uvital , v Vvital , w Wvital }. We denote the vital phases Pvital (u, v, w) and non-vital phases Pnonvital (u, v, w) as follows: P(u, v, w) = Pvital(u, v, w) if (u, v, w) Cvital Pnonvital(u, v, w) otherwise. (4) Pvital (u, v, w) and Pnonvital (u, v, w) are classified based on robustness-related weights within a specified frequency range. The filter with a specific frequency range can prevent an increase in phase feature loss in that frequency range by not applying the filter to the entire phases at once. Vital Phase Augmentation To ensure that the model depends on the vital phases, we apply weak variations to the vital phases, whereas strong variations are applied to the non-vital phases. Excessive variations impede the model from learning domain-invariant features from the phase spectrum. There are two types of phase augmentation: vital phase augmentation using a Gaussian distribution (VIPAug-G) and using fractal phases (VIPAug F) VIPAug-G. VIPAug-G involves random sampling from a zero-mean Gaussian distribution and adding the obtained values to the phases. A Gaussian distribution exhibits high probability density around the mean and low probability density away from the mean. A Gaussian distribution effectively introduces finite variations to the phases to strengthen the model s dependency on the vital phases. The variations from Gaussian distributions N with different variances are applied based on the corresponding weights: P gauss aug (u, v, w) = P(u, v, w) + V (u, v, w), (5) where a random variable V (u, v, w) N 0, σ2 vital if (u, v, w) Cvital, V (u, v, w) N 0, σ2 nonvital otherwise and σ2 vital σ2 nonvital . With π Pvital π and π Pnonvital π, the variation should be correspondingly small for the narrow range of vital and non-vital phases. In contrast to pixel-level perturbation caused by Gaussian noise, VIPAug-G introduces variations to the phases of the complex exponential functions that comprise an image by linear combination. VIPAug-F. VIPAug-F preserves the vital phases and entirely substitutes the non-vital phases with fractal phases. This method induces larger variations than VIPAug-G to enhance robustness against more significant fluctuations in phases. The replacement images should be from another domain and that domain should have different classes compared with the source domain. Replacing the original phases with those from other images from the same source domain prevents the model from depending on the phases, and diminishes the model s capability to learn domain-invariant features from the phases. Hence, we use the phases of a fractal image to enhance the structural complexity (Hendrycks et al. 2022) of the image. Fractal images are randomly chosen from a pool of 14,200 images. The non-vital phases are replaced with the fractal phases: P frac aug (u, v, w) = P(u, v, w) if (u, v, w) Cvital Pfractal(u, v, w) otherwise, (6) where Pfractal (u, v, w) denotes phases of the fractal image. VIPAug-F is designed to be robust against stronger phase fluctuations. However, completely replacing the non-vital phases may result in a substantial loss of image features. To retain original image features in the non-vital phases, we randomly apply VIPAug-F each iteration. Additionally, optional modifications may be necessary according to the training dataset. Due to more image features falling into the lowfrequency region (Li et al. 2023), the non-vital phase with the highest weight is retained at the low-frequency region. This is because the non-vital phases also have different relative robustness weights depending on the magnitude of the amplitudes. VIPAug. VIPAug combines amplitude augmentation and two types of phase augmentation. Denoting VIPAug-G as function g( ) and VIPAug-F as function h( ), we obtain Paug = g t h(P), (7) where denotes function composition and t( ) is a phase change by pixel-wise augmentations from Auto Augment. The augmented amplitude Aaug is obtained by APR-SP. We can reconstruct the augmented image faug through i DFT with Paug and Aaug. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Dataset Method Clean Acc (%) Corruption error rate (%) Baseline 95.3 25.3 APR-SP 95.6 8.7 PRIME 93.9 10.1 VIPAug-G 95.6 8.4 VIPAug-F 95.8 8.6 VIPAug 95.8 8.4 Baseline 78.3 51.3 APR-SP 76.7 31.5 PRIME 76.8 31.9 VIPAug-G 76.9 31.2 VIPAug-F 77.3 31.3 VIPAug 77.2 31.2 Table 1. Comparison with state-of-the-art methods on CIFAR-10 and CIFAR-100. The clean accuracy and corruption error rate were evaluated. Experiments Datasets. We experimentally evaluated the performance of VIPAug on the most widely used CIFAR-10, CIFAR100, Image Net-100, and Image Net datasets. CIFAR-10 and CIFAR-100 comprise 50,000 training images and 10,000 testing images, and each image is a 32 32 color image with 10 classes and 100 classes, respectively. Image Net consists of 1.2 million images and 1,000 classes. Image Net-100 consists of 100 randomly selected classes of Image Net. The training and test dataset contain 1,300 and 50 images per class, respectively. We used 14,200 fractal images from collections on Deviant Art to train the model. To measure the domain generalization performance, we used the corrupted datasets CIFAR-10-C, CIFAR-100-C, Image Net-100-C, and Image Net-C (Hendrycks and Dietterich 2019), which contain 15 types of corruption, including noise, blur, weather, and digital corruption. Each type is demonstrated at five levels of severity. Metrics. We evaluated the domain generalization performance of the proposed method by measuring its accuracy on clean images and classification error rates on corrupted images. We also used the mean corruption error (m CE) on Image Net-100-C and Image Net-C, which is a normalized measure of the classification error rate by using Alex Net (Krizhevsky, Sutskever, and Hinton 2012). The corrupted test data has five severity levels 1 s 5. The corruption error for each type of corruption was calculated as follows: CENetwork Corruption = P5 s=1 ENetwork s,Corruption/ P5 s=1 EAlex Net s,Corruption. We then calculate the m CE by averaging the CENetwork Corruption for each type of corruption. CIFAR-10 and CIFAR-100 Training Setup. We used a Res Net-18 (He et al. 2016) architecture as a baseline model. We trained all methods for Method Clean Acc (%) Corruption error rate (%) m CE (%) Baseline 81.8 52.8 82.6 APR-SP 82.2 41.0 65.4 PRIME 80.2 38.3 60.9 VIPAug-G 82.1 39.5 63.4 VIPAug-F 82.4 40.2 64.5 VIPAug 82.3 39.4 63.2 VIPAug with color 80.5 38.3 61.0 VIPAug + PRIME 79.8 34.3 55.4 Table 2. Comparison with state-of-the-art methods on Image Net-100. Clean accuracy, corruption error rate, and mean corruption error (m CE) were evaluated. 250 epochs. Detailed training setup can be seen in the supplementary material. We used the 2 2 1 argmax filter, and set σvital = 0.001 and σnonvital = 0.014 on CIFAR-10 and σvital = 0.005 and σnonvital = 0.012 on CIFAR-100. VIPAug-G uses small values for the variance to introduce small variation to the phase. More details can be seen in the supplementary material. We also applied the modification to VIPAug-F on CIFAR-10 by setting the low-frequency region to 4/9 of the total phase. The non-vital phase with the highest weight is retained at the low frequency region. Results. Table 1 shows the performance comparison with the state-of-the-art models on CIFAR-10 and CIFAR-100. The baseline model achieved 95.3% accuracy on the clean domain of CIFAR-10. However, the performance dropped significantly with an error rate of 25.3% in the corrupted domain, which shows the importance of domain generalization. PRIME improved performance in the corrupted domain with primitive augmentations, but the method suffered from performance degradation in the clean domain. APR-SP fixes the phases and replaces the amplitudes with the amplitudes of other images. APR-SP improved accuracy in both the clean and corrupted domains. However, the method only considers the amplitude replacement, making it vulnerable to phase fluctuations in common corruptions. VIPAug combines phase variations with amplitude replacement to perform robustly against phase and amplitude fluctuations. VIPAug achieved an accuracy of 95.8% and a corruption error rate of 8.4%. Compared with APR-SP, these values were 0.2%p and 0.3%p better, respectively. VIPAug outperformed all the other methods on both clean and corrupted datasets. VIPAug-G and VIPAug achieved the lowest error rate for corrupted data, while VIPAug-F and VIPAug achieved the highest accuracy for clean data. Compared with APR-SP, VIPAug-G outperformed on the corrupted domain, indicating that VIPAug-G is more robust to corruption without sacrificing accuracy on uncorrupted data. VIPAug-F also exhibited improved performance for clean and corrupted data compared with APR-SP, despite the strong variation. These results show that the model still learns domain-invariant features from the phase, even with the high variation of VIPAug-F. VIPAug achieved state-of- The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Noise Blur Weather Digital Method Clean Gauss. Shot Impulse Defocus Glass Motion Zoom Snow Frost Fog Bright Contrast Elastic Pixel JPEG m CE Baseline 23.9 79 80 82 82 90 84 80 86 81 75 65 79 91 77 80 80.6 Patch Uniform 24.5 67 68 70 74 83 81 77 80 74 75 62 77 84 71 71 74.3 Auto Aug (AA) 22.8 69 68 72 77 83 80 81 79 75 64 56 70 88 57 71 72.7 Random AA 23.6 70 71 72 80 86 82 81 81 77 72 61 75 88 73 72 76.1 Max Blur pool 23.0 73 74 76 74 86 78 77 77 72 63 56 68 86 71 71 73.4 SIN 27.2 69 70 70 77 84 76 82 74 75 69 65 69 80 64 77 73.3 Aug Mix 22.4 65 66 67 70 80 66 66 75 72 67 58 58 79 69 69 68.4 APR-SP 24.4 60 64 63 70 85 69 80 68 68 56 56 63 81 65 63 67.4 VIPAug (Ours) 24.1 56 59 57 70 84 69 79 64 64 55 55 65 81 63 67 65.8 Table 3. Comparison with state-of-the-art methods on Image Net dataset. For each corruption type, the average performance was evaluated. The mean corruption error (m CE) is a normalized average measure of the classification error rate on 15 corruptions. the-art performance on CIFAR-10 and CIFAR-100. Image Net-100 and Image Net Training Setup. We used a Res Net-18 architecture as the baseline model on Image Net-100 and a Res Net-50 (He et al. 2016) model on Image Net. The models were trained for 100 epochs. Detailed training setup can be seen in the supplementary material. We evaluated all methods on Image Net100 using the same training settings. For Image Net, we used pretrained weights for alternative methods if available. Otherwise, we used the performance results reported in Aug Mix. We used the 2 2 1 argmax filter, and set σvital = 0.001 and σnonvital = 0.005. We applied the modification to VIPAug-F by setting the low-frequency region to 1/4 of the total phase. Results. In Table 2, we compared VIPAug with APR-SP and PRIME on Image Net-100. The variations of VIPAug exhibited greater accuracy on the clean domain compared with the other methods. VIPAug also showed greater clean accuracy by 0.1%p and decreased m CE by 2.2%p compared with APR-SP. PRIME reduced the clean accuracy compared with the baseline, but significantly improved the generalization ability on the corrupted domain. We conjecture that the diverse color transformation of PRIME contributed to the generalization capability. We added the color transformation of PRIME to VIPAug and compared the performance. VIPAug with color decreased the m CE by 4.4%p compared with APR-SP and achieved nearly state-of-the-art performance compared with PRIME. When VIPAug and PRIME were applied together, they showed an overwhelming performance of 55.4% m CE. These results confirm that VIPAug and PRIME can be considered somewhat orthogonal approaches. In Table 3, domain generalization methods were evaluated on Image Net for each corruption type (Lopes et al. 2019; Cubuk et al. 2019; Zhang 2019; Rusak et al. 2020; Hendrycks et al. 2019; Chen et al. 2021). VIPAug did not excel in all corruption types, but the method achieved the best average performance on the corrupted domains while maintaining performance on the clean domain. This demonstrates the effectiveness of VIPAug on a large-scale dataset. Method Clean Acc (%) Corruption error rate (%) m CE (%) APR-SP 82.2 41.0 65.4 VIPAug 82.3 39.4 63.2 Reverse VIPAug 81.8 42.8 68.2 Uniform VIPAug (a) 82.2 40.5 64.9 Uniform VIPAug (b) 81.7 41.2 66.1 Table 4. The ablation analysis of robustness weight to verify two hypotheses. VIPAug outperformed Reverse VIPAug, implying that vital phase contains more robust features. Furthermore, VIPAug outperformed Uniform VIPAug (a) and (b), implying that strengths of variation should be proportional to the robustness of each phase. Ablation Studies Robustness Weight. We evaluated our first hypothesis that vital phase contains more domain-invariant features than non-vital phase by comparing the performance of Reverse VIPAug and VIPAug. Reverse VIPAug treated vital phase as non-vital phase and one of non-vital phases as vital phase. If the robustness weights of vital phase and non-vital phase are the same, the performance should be similar. However, as shown in Table 4, Reverse VIPAug performed worse than VIPAug. The clean accuracy of Reverse VIPAug was 0.5%p lower than that of VIPAug, and the m CE was 5.0%p higher. This indicates that Reverse VIPAug is not as robust to corruption as VIPAug. In particular, when compared with APR-SP without phase variation, the clean accuracy of Reverse VIPAug was 0.4%p lower and the m CE was 2.8%p higher. This suggests that adding variation to the phase does not always improve performance. We then evaluated the second hypothesis that the strengths of variation should be proportional to the robustness weight of each phase by comparing the performance of Uniform VIPAug (a) and Uniform VIPAug (b). Uniform VIPAug (a) set σvital = 0.001 and σnonvital = 0.001, and randomly replaced vital phase and non-vital phase with fractals to give the same strengths of variation. Uniform VIPAug (b) set σvital = 0.005 and σnonvital = 0.005, with other conditions The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 5. The ablation analysis of modifications to the phase range of VIPAug-F on Image Net-100 and Image Net-100-C. The x-axis stands for the ratio of the modified phase range to the total phase spectrum. Dataset Clean Acc (%) Corruption error rate (%) Fractal 77.3 31.4 Image Net (IN) 77.0 31.6 Stylized-IN 77.1 31.7 GTA5 76.9 31.5 Table 5. Ablation analysis of VIPAug-F for other datasets. Only fractal dataset has no class. same as Uniform VIPAug (a). Uniform VIPAug (a) and (b) exhibited lower performance on both clean and corrupted images compared with VIPAug. Our results suggest that the model becomes more robust to corruption if the strengths of variations are varied according to the robustness weights. Modification on VIPAug-F. VIPAug-F introduces strong variations to partially replace phases with fractal images phases. To investigate the effects of different modification ranges on the performance of VIPAug-F, we conducted an ablation study on Image Net-100. Because the robustness weights are relatively different between non-vital phases, we modified VIPAug-F not to additionally replace the case with the second largest magnitude of amplitude at the low-frequency spectrum. This is because the low-frequency region contains more image features (Li et al. 2023). We conducted the experiment in four cases: no modification, modification of 1/4 of the entire phases, modification of 4/9, and modification of the entire phases, as shown in Figure 5. We found that VIPAug-F with no modification had the lowest clean accuracy and the highest m CE. VIPAug-F with more modification had higher clean accuracy but lower m CE. These results suggest that finite variation should be given to the phase in order to improve the performance of VIPAug-F empirically. We need to find the appropriate hyperparameters that balance the clean accuracy and m CE. Other Datasets for VIPAug-F. In Table 5, we compared the performance of VIPAug-F when using different datasets instead of fractals. We used Image Net, Stylized-Image Net, and GTA5 (Richter et al. 2016) datasets. The number of im- Figure 6. Comparison of applying 2D DFT and 3D DFT to VIPAug on CIFAR-100. ages was 14,200, the same as fractal images. The 14,200 images were randomly selected from each dataset. Other datasets show a slight performance decrease for clean and corrupted datasets compared with fractal dataset. This is because Image Net, Stylized-Image Net, and GTA5 all have similar classes to the CIFAR-100. If two images of different classes are mixed together, the model cannot depend on the phase well. On the other hand, the fractal images have no class. Fractals also introduce structural complexity to images (Hendrycks et al. 2022). CIFAR-100 has a wide variety of classes, making it difficult to find a dataset consisting of completely different images. Therefore, the dataset without a class is more suitable. Comparison with 2D DFT. We compared the performance of VIPAug and 2D-DFT VIPAug in Figure 6. 2DDFT VIPAug showed 0.7%p lower clean accuracy and 2.3%p higher corruption error rate than VIPAug. Extending along the channel axis in 3D DFT allows vital phases to be identified for all channels when the filter is applied. Considering the large performance difference, amplitude and phase features between channels have a significant impact on the model s performance. Conclusion We first argue that the robustness weight differs for each phase of the image. We propose a novel method to classify vital and non-vital phases according to their weights. We make the model more robust to corruption by giving different strengths of variation to the phases according to their weights. Our extensive experimental results showed that our approach achieved SOTA performance on CIFAR-10 and CIFAR-100, and achieved performance close to the SOTA methods on Image Net-100 and Image Net. In this study, we presented a new perspective on the phase in domain generalization research. We suggested a promising direction for subsequent research on how to deal with the image phase. Acknowledgments This work was supported in part by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by Korea government (MSIT) (No.2020-0-00440, Development of Artificial Intelligence Technology that Continuously Improves Itself as the The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Situation Changes in the Real World). This research was supported in part by the KAIST Convergence Research Institute Operation Program. The students are supported by the BK21 FOUR from the Ministry of Education (Republic of Korea). References Chen, G.; Peng, P.; Ma, L.; Li, J.; Du, L.; and Tian, Y. 2021. Amplitude-phase recombination: Rethinking robustness of convolutional neural networks in frequency domain. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 458 467. Cubuk, E. D.; Zoph, B.; Mane, D.; Vasudevan, V.; and Le, Q. V. 2019. Auto Augment: Learning augmentation policies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 113 123. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei Fei, L. 2009. Image Net: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 248 255. De Vries, T.; and Taylor, G. W. 2017. Improved regularization of convolutional neural networks with cutout. ar Xiv:1708.04552. Ding, Z.; and Fu, Y. 2017. Deep domain generalization with structured low-rank constraint. IEEE Transactions on Image Processing, 27(1): 304 313. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; Uszkoreit, J.; and Houlsby, N. 2021. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations. He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 770 778. Hendrycks, D.; and Dietterich, T. 2019. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations. Hendrycks, D.; Mu, N.; Cubuk, E. D.; Zoph, B.; Gilmer, J.; and Lakshminarayanan, B. 2019. Aug Mix: A simple data processing method to improve robustness and Uncertainty. In International Conference on Learning Representations. Hendrycks, D.; Zou, A.; Mazeika, M.; Tang, L.; Li, B.; Song, D.; and Steinhardt, J. 2022. Pix Mix: Dreamlike pictures comprehensively improve safety measures. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 16783 16792. Krizhevsky, A.; and Hinton, G. 2009. Learning multiple layers of features from tiny images. Technical Report TR-2009, Dept. of Computer Science, Toronto Univ. Krizhevsky, A.; Sutskever, I.; and Hinton, G. E. 2012. Image Net classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25. Lee, W.; and Myung, H. 2022. Adversarial attack for asynchronous event-based data. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, 1237 1244. Li, Z.; Ortega Caro, J.; Rusak, E.; Brendel, W.; Bethge, M.; Anselmi, F.; Patel, A. B.; Tolias, A. S.; and Pitkow, X. 2023. Robust deep learning object recognition models rely on low frequency information in natural images. PLOS Computational Biology, 19(3): e1010932. Liu, Q.; Dou, Q.; Yu, L.; and Heng, P. A. 2020. MS-Net: multi-site network for improving prostate segmentation with heterogeneous MRI data. IEEE Transactions on Medical Imaging, 39(9): 2713 2724. Lopes, R. G.; Yin, D.; Poole, B.; Gilmer, J.; and Cubuk, E. D. 2019. Improving robustness without sacrificing accuracy with patch gaussian augmentation. ar Xiv:1906.02611. Modas, A.; Rade, R.; Ortiz-Jim enez, G.; Moosavi-Dezfooli, S.-M.; and Frossard, P. 2022. PRIME: A few primitives can boost robustness to common corruptions. In Proceedings of the European Conference on Computer Vision, 623 640. Springer. Motiian, S.; Piccirilli, M.; Adjeroh, D. A.; and Doretto, G. 2017. Unified deep supervised domain adaptation and generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 5715 5725. Richter, S. R.; Vineet, V.; Roth, S.; and Koltun, V. 2016. Playing for data: Ground truth from computer games. In Proceedings of the European Conference on Computer Vision, 102 118. Springer. Rusak, E.; Schott, L.; Zimmermann, R. S.; Bitterwolf, J.; Bringmann, O.; Bethge, M.; and Brendel, W. 2020. A simple way to make neural networks robust against diverse image corruptions. In Proceedings of the European Conference on Computer Vision, 53 69. Springer. Tan, M.; and Le, Q. 2019. Efficient Net: Rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning, 6105 6114. PMLR. Xu, Q.; Zhang, R.; Zhang, Y.; Wang, Y.; and Tian, Q. 2021. A fourier-based framework for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 14383 14392. Yoon, C.; Hamarneh, G.; and Garbi, R. 2019. Generalizable feature learning in the presence of data bias and domain class imbalance with application to skin lesion classification. In International Conference on Medical Image Computing and Computer Assisted Intervention, 365 373. Springer. Yucel, M. K.; Cinbis, R. G.; and Duygulu, P. 2023. Hybrid Augment++: Unified frequency spectra perturbations for model robustness. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 5718 5728. Zhang, H.; Cisse, M.; Dauphin, Y. N.; and Lopez-Paz, D. 2018. Mixup: Beyond empirical risk minimization. In International Conference on Learning Representations. Zhang, R. 2019. Making convolutional networks shiftinvariant again. In International Conference on Machine Learning, 7324 7334. PMLR. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Zhao, Y.; Zhong, Z.; Yang, F.; Luo, Z.; Lin, Y.; Li, S.; and Sebe, N. 2021. Learning to generalize unseen domains via memory-based multi-source meta-learning for person reidentification. In Proceedings of the IEEE/CVF International Conference on Computer Vision and Pattern Recognition, 6277 6286. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; and Yang, Y. 2020. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, 13001 13008. Zhou, K.; Yang, Y.; Qiao, Y.; and Xiang, T. 2021. Domain generalization with mixstyle. In International Conference on Learning Representations. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)