# improved_transformer_for_highresolution_gans__1504d9e8.pdf Improved Transformer for High-Resolution GANs Long Zhao1, Zizhao Zhang2 Ting Chen3 Dimitris N. Metaxas1 Han Zhang3 1Rutgers University 2Google Cloud AI 3Google Research Attention-based models, exemplified by the Transformer, can effectively model long range dependency, but suffer from the quadratic complexity of self-attention operation, making them difficult to be adopted for high-resolution image generation based on Generative Adversarial Networks (GANs). In this paper, we introduce two key ingredients to Transformer to address this challenge. First, in low-resolution stages of the generative process, standard global self-attention is replaced with the proposed multi-axis blocked self-attention which allows efficient mixing of local and global attention. Second, in high-resolution stages, we drop self-attention while only keeping multi-layer perceptrons reminiscent of the implicit neural function. To further improve the performance, we introduce an additional self-modulation component based on cross-attention. The resulting model, denoted as Hi T, has a nearly linear computational complexity with respect to the image size and thus directly scales to synthesizing high definition images. We show in the experiments that the proposed Hi T achieves state-of-the-art FID scores of 30.83 and 2.95 on unconditional Image Net 128 128 and FFHQ 256 256, respectively, with a reasonable throughput. We believe the proposed Hi T is an important milestone for generators in GANs which are completely free of convolutions. Our code is made publicly available at https://github.com/google-research/hit-gan. 1 Introduction Attention-based models demonstrate notable learning capabilities for both encoder-based and decoderbased architectures [66, 69] due to their self-attention operations which can capture long-range dependencies in data. Recently, Vision Transformer [12], one of the most powerful attention-based models, has achieved a great success on encoder-based vision tasks, specifically image classification [12, 60], segmentation [37, 64], and vision-language modeling [46]. However, applying the Transformer to image generation based on Generative Adversarial Networks (GANs) is still an open problem. The main challenge of adopting the Transformer as a decoder/generator lies in two aspects. On one hand, the quadratic scaling problem brought by the self-attention operation becomes even worse when generating pixel-level details for high-resolution images. For example, synthesizing a high definition image with the resolution of 1024 1024 leads to a sequence containing around one million pixels in the final stage, which is unaffordable for the standard self-attention mechanism. On the other hand, generating images from noise inputs poses a higher demand for spatial coherency in structure, color, and texture than discriminative tasks, and hence a more powerful yet efficient self-attention mechanism is desired for decoding feature representations from inputs. In view of these two key challenges, we propose a novel Transformer-based decoder/generator in GANs for high-resolution image generation, denoted as Hi T. Hi T employs a hierarchical structure of Transformers and divides the generative process into low-resolution and high-resolution stages, focusing on feature decoding and pixel-level generating, respectively. Specifically, its low-resolution This work was done while Long Zhao was a student researcher at the Google Brain team. Correspondence to: Long Zhao (lz311@rutgers.edu) and Han Zhang (zhanghan@google.com). 35th Conference on Neural Information Processing Systems (Neur IPS 2021). stages follow the design of Nested Transformers [73] but enhanced by the proposed multi-axis blocked self-attention to better capture global information. Assuming that spatial features are well decoded after low-resolution stages, in high-resolution stages, we drop all self-attention operations in order to handle extremely long sequences for high definition image generation. The resulting high-resolution stages of Hi T are built by multi-layer perceptrons (MLPs) which have a linear complexity with respect to the sequence length. Note that this design aligns with the recent findings [58, 59] that pure MLPs manage to learn favorable features for images, but it simply reduces to an implicit neural function [41, 42, 43] in the case of generative modeling. To further improve the performance, we present an additional cross-attention module that acts as a form of self-modulation [8]. In summary, this paper makes the following contributions: We propose Hi T, a Transformer-based generator for high-fidelity image generation. Standard self-attention operations are removed in the high-resolution stages of Hi T, reducing them to an implicit neural function. The resulting architecture easily scales to high definition image synthesis (with the resolution of 1024 1024) and has a comparable throughput to Style GAN2 [28]. To tame the quadratic complexity and enhance the representation capability of self-attention operation in low-resolution stages, we present a new form of sparse self-attention operation, namely multi-axis blocked self-attention. It captures local and global dependencies within nonoverlapping image blocks in parallel by attending to a single axis of the input tensor at a time, each of which uses a half of attention heads. The proposed multi-axis blocked self-attention is efficient, simple to implement, and yields better performance than other popular self-attention operations [37, 62, 73] working on image blocks for generative tasks. In addition, we introduce a cross-attention module performing attention between the input and intermediate features. This module re-weights intermediate features of the model as a function of the input, playing a role as self-modulation [8], and provides important global information to high-resolution stages where self-attention operations are absent. The proposed Hi T obtains competitive FID [16] scores of 30.83 and 2.95 on unconditional Image Net [49] 128 128 and FFHQ [28] 256 256, respectively, highly reducing the gap between Conv Net-based GANs and Transformer-based ones. We also show that Hi T not only works for GANs but also can serve as a general decoder for other models such as VQ-VAE [61]. Moreover, we observe that the proposed Hi T can obtain more performance improvement from regularization than its Conv Net-based counterparts. To the best of our knowledge, these are the best reported scores for an image generator that is completely free of convolutions, which is an important milestone towards adopting Transformers for high-resolution generative modeling. 2 Related work Transformers for image generation. There are two main streams of image generation models built on Transformers [63] in the literature. One stream of them [14, 44] is inspired by auto-regressive models that learn the joint probability of the image pixels. The other stream focuses on designing Transformer-based architecture for generative adversarial networks (GANs) [15]. This work follows the spirit of the second stream. GANs have made great progress in various image generation tasks, such as image-to-image translation [57, 74, 75, 79] and text-to-image synthesis [70, 71], but most of them depend on Conv Net-based backbones. Recent attempts [23, 32] build a pure Transformer-based GAN by a careful design of attention hyper-parameters as well as upsampling layers. However, such a model is only validated on small scale datasets (e.g., CIFAR-10 [31] and STL-10 [10] consisting of 32 32 images) and does not scale to complex real-world data. To our best knowledge, no existing work has successfully applied a Transformer-based architecture completely free of convolutions for high-resolution image generation in GANs. GANformer [21] leverages the Transformer as a plugin component to build the bipartite structure to allow long-range interactions during the generative process, but its main backbone is still a Conv Net based on Style GAN [28]. GANformer and Hi T are different in the goal of using attention modules. GANformer utilizes the attention mechanism to model the dependences of objects in a generated scene/image. Instead, Hi T explores building efficient attention modules for synthesizing general objects. Our experiments demonstrate that even for simple face image datasets, incorporating the attention mechanism can still lead to performance improvement. We believe our work reconfirms the necessity of using attention modules for general image generation tasks, and more importantly, we Position Encoding Up Sample K V Position Encoding Latent Embedding Low-Resolution Stages: Multi-Axis Nested Transformer High-Resolution Stages: Implicit Functions Input Latent Code Output Image Latent Embedding Latent Embedding B1 B0 B8 B7 Linear Projection Figure 1: Hi T architecture. In each stage, input the feature is first organized into blocks (denoted as Bi). Hi T s low-resolution stages follow the decoder design of Nested Transformer [73] which is then enhanced by the proposed multi-axis blocked self-attention. We drop self-attention operations in the high-resolution stages, resulting in implicit neural functions. The model is further boosted by cross-attention modules which allow intermediate features to be modulated directly by the input latent code. The detailed algorithm can be found in the supplementary materials. present an efficient architecture for attention-based high-resolution generators which might benefit the design of future attention-based GANs. Attention models. Many efforts [2, 17, 65] have been made to reduce the quadratic complexity of self-attention operation. When handling vision tasks, some works [37, 62, 73] propose to compute local self-attention in blocked images which takes advantage of the grid structure of images. [6] also presents local self-attention within image blocks but it does not perform self-attention across blocks as in Hi T. The most related works to the proposed multi-axis blocked self-attention are [17, 65] where they also compute attention along axes. But our work differs notably in that we compute different attentions within heads on blocked images. Some other works [7, 21, 22] avoid directly applying standard self-attention to the input pixels, and perform attention between the input and a small set of latent units. Our cross-attention module differs from them in that we apply cross-attention to generative modeling designed for pure Transformer-based architectures. Implicit neural representations. The largest popularity of implicit neural representations/functions (INRs) is studied in 3D deep learning to represent a 3D shape in a cheap and continuous way [41, 42, 43]. Recent studies [1, 13, 34, 54] explore the idea of using INRs for image generation, where they learn a hyper MLP network to predict an RGB pixel value given its coordinates on the image grid. Among them, [1, 54] are closely related to our high-resolution stages of the generative process. One remarkable difference is that our model is driven by the cross-attention module and features generated in previous stages instead of the hyper-network presented in [1, 54]. 3.1 Main Architecture In the case of unconditional image generation, Hi T takes a latent code z N(0, I) as input and generates an image of the target resolution through a hierarchical structure. The latent code is first projected into an initial feature with the spatial dimension of H0 W0 and channel dimension of C0. During the generation process, we gradually increase the spatial dimension of the feature map while reducing the channel dimension in multiple stages. We divide the generation stages into low-resolution stages and high-resolution stages to balance feature dependency range in decoding and computation efficiency. The overview of the proposed method is illustrated in Figure 1. In low-resolution stages, we allow spatial mixing of information by efficient attention mechanism. We follow the decoder form of Nested Transformer [73] where in each stage, the input feature is first [4, 4, C / 2] [4, 4, C / 2] Attention on 1st Axis Input [4, 4, C] Blocked Input [2 x 2, 2 x 2, C] Output [4, 4, C] Concat & Unblocking Channel Splitting 1 / 2 Heads [4, 4, C / 2] 1 / 2 Heads [4, 4, C / 2] Attention on 2nd Axis Figure 2: Multi-axis self-attention architecture. The different stages of multi-axis self-attention for a [4, 4, C] input with the block size of b = 2. The input is first blocked into 2 2 non-overlapping [2, 2, C] patches. Then regional and dilated self-attention operations are computed along two different axes, respectively, each of which uses a half of attention heads. The attention operations run in parallel for each of the tokens and their corresponding attention regions, illustrated with different colors. The spatial dimensions after attention are the same as the input image. divided into non-overlapping blocks where each block can be considered as a local patch. After being combined with learnable positional encoding, each block is processed independently via a shared attention module. We enhance the local self-attention of Nested Transformer by the proposed multiaxis blocked self-attention that can produce a richer feature representation by explicitly considering local (within blocks) as well as global (across blocks) relations. We denote the overall architecture of these stages as multi-axis Nested Transformer. Assuming that spatial dependency is well modeled in the low-resolution stages, the high-resolution stages can focus on synthesizing pixel-level image details purely based on the local features. Thus in the high-resolution stage, we remove all self-attention modules and only maintain MLPs which can further reduce computation complexity. The resulting architecture in this stage can be viewed as an implicit neural function conditioned on the given latent feature as well as positional information. We further enhance Hi T by adding a cross-attention module at the beginning of each stage which allows the network to directly condition the intermediate features on the initial input latent code. This kind of self-modulation layer leads to improved generative performance, especially when selfattention modules are absent in high-resolution stages. In the following sections, we provided detailed descriptions for the two main architectural components of Hi T: (i) multi-axis blocked self-attention module and (ii) cross-attention module, respectively. 3.2 Multi-Axis Blocked Self-Attention Different from the blocked self-attention in Nested Transformer [73], the proposed multi-axis blocked self-attention performs attention on more than a single axis. The attentions performed in two axes correspond to two forms of sparse self-attention, namely regional attention and dilated attention. Regional attention follows the spirit of blocked local self-attention [37, 62] where tokens attend to their neighbours within non-overlapped blocks. To remedy the loss of global attention, dilated attention captures long-range dependencies between tokens across blocks: it subsamples attention regions in a manner similar to dilated convolutions with a fixed stride equal to the block size. Figure 2 illustrates an example of these two attentions. To be specific, given an input of image with size (H, W, C), it is blocked into a tensor X of the shape (b b, H b , C) representing (b, b) non-overlapping blocks each with the size of ( H b ). Dilated attention mixes information along the first axis of X while keeping information along other axes independent; regional attention works in an analogous manner over the second axis of X. They are straightforward to implement: attention over the i-th axis of X can be implemented by einsum operation which is available in most deep learning libraries. We mix regional and dilated attention in a single layer by computing them in parallel, each of which uses a half of attention heads. Our method can be easily extended to model more than two axes by performing attention on each of the axis in parallel. Axial attention [17, 20, 65] can be viewed as a special case of our method, where the blocking operation before attention is removed. However, we find blocking is the key to achieve significant performance improvement in the experiment. This is because the blocked structure reveals a better inductive bias for images. Compared with [17, 37, 65] where different attention modules are interleaved in consecutive layers, our approach aggregates local and global information in a single round, which is not only more flexible for architecture design but also shown to yield better performance than interleaving in our experiment. Balancing attention between axes. In each multi-axis blocked self-attention module, the input feature is blocked in a balanced way such that we have b b H b . This ensures that regional and dilated attention is computed on an input sequence of a similar length, avoiding half of the attention heads are attended to a too sparse region. In general, performing dot-product attention between two input sequences of length N = H W requires a total of O(N 2) computation. When computing the balanced multi-axis blocked attention on an image with the block size of S, i.e., S = N, we perform attention on S sequences of length S, which is a total of O(S S2) = O(N N) computation. This leads to an O( N) saving in computation over standard self-attention. 3.3 Cross-Attention for Self-Modulation To further improve the global information flow, we propose to let the intermediate features of the model directly attend to a small tensor projected from the input latent code. This is implemented via a cross-attention operation and can be viewed as a form of self-modulation [8]. The proposed technique has two benefits. First, as shown in [8], self-modulation stabilizes the generator towards favorable conditioning values and also appears to improve mode coverage. Second, when self-attention modules are absent in high-resolution stages, attending to the input latent code provides an alternative way to capture global information when generating pixel-level details. Formally, let Xl be the first-layer feature representation of the l-th stage. The input latent code z is first projected into a 2D spatial embedding Z with the resolution of HZ WZ and dimension of CZ by a linear function. Xl is then treated as the query and Z as the key and value. We compute their cross-attention following the update rule: X l = MHA(Xl, Z + PZ), where MHA represents the standard mulit-head self-attention, X l is the output, and PZ is the learnable positional encoding having the same shape as Z. Note that Z is shared across all stages. For an input feature with the sequence length of N, the embedding size is a pre-defined hyperparameter far less than N (i.e., HZ WZ N). Therefore, the resulting cross-attention operation has linear complexity O(N). In our initial experiments, we find that compared with cross-attention, using Ada IN [19] and modulated layers [28] for Transformer-based generators requires much higher memory cost during model training, which usually leads to out-of-memory errors when the model is trained for generating highresolution images. As a result, related work like Vi T-GAN [32], which uses Ada IN and modulated layers, can only produce images up to the resolution of 64 64. Hence, we believe cross-attention is a better choice for high-resolution generators based on Transformers. 4 Experiments 4.1 Experiment Setup Datasets. We validate the proposed method on three datasets: Image Net [49], Celeb A-HQ [25], and FFHQ [28]. Image Net (LSVRC2012) [49] contains roughly 1.2 million images with 1000 distinct categories and we down-sample the images to 128 128 and 256 256 by bicubic interpolation. We use random crop for training and center crop for testing. This dataset is challenging for image generation since it contains samples with diverse object categories and textures. We also adopt Image Net as the main test bed during the ablation study. Celeb A-HQ [25] is a high-quality version of the Celeb A dataset [38] containing 30,000 of the facial images at 1024 1024 resolution. To align with [25], we use these images for both training and evaluation. FFHQ [28] includes vastly more variation than Celeb A-HQ in terms of age, ethnicity and image background, and also has much better coverage of accessories such as eyeglasses, sunglasses, and hats. This dataset consists of 70,000 high-quality images at 1024 1024 resolution, out of which we use 50,000 images for testing and train models with all images following [28]. We synthesize images on these two datasets with the resolutions of 256 256 and 1024 1024. Evaluation metrics. We adopt the Inception Score (IS) [51] and the Fréchet Inception Distance (FID) [16] for quantitative evaluation. Both metrics are calculated based on a pre-trained Inception-v3 image classifier [56]. Inception score computes KL-divergence between the real image distribution and the generated image distribution given the pre-trained classifier. Higher inception scores mean better image quality. FID is a more principled and comprehensive metric, and has been shown to be more consistent with human judgments of realism [16, 70]. Lower FID values indicate closer distances between synthetic and real data distributions. In our experiments, 50,000 samples are randomly generated for each model to calculate the inception score and FID on Image Net and FFHQ, while 30,000 samples are produced for comparison on Celeb A-HQ. Note that we follow [51] to split the synthetic images into groups (5000 images per group) and report their averaged inception score. 4.2 Implementation Details Architecture configuration. In our implementation, Hi T starts from an initial feature of size 8 8 projected from the input latent code. We use pixel shuffle [53] for upsampling the output of each stage, as we find using nearest neighbors leads to model failure which aligns with the observation in Nested Transformer [73]. The number of low-resolution stages is fixed to be 4 as a good trade-off between computational speed and generative performance. For generating images larger than the resolution of 128 128, we scale Hi T to different model capacities small, base, and large, denoted as Hi T-{S, B, L}. We refer to the supplementary materials for their detailed architectures. It is worth emphasizing that the flexibility of the proposed multi-axis blocked self-attention makes it possible for us to build smaller models than [37]. This is because they interleave different types of attention operations and thus require the number of attention blocks in a single stage to be even (at least 2). In contrast, our multi-axis blocked self-attention combines different attention outputs within heads which allows us to use an arbitrary number (e.g., 1) of attention blocks in a model. Training details. In all the experiments, we use a Res Net-based discriminator following the architecture design of [28]. Our model is trained with a standard non-saturating logistic GAN loss with R1 gradient penalty [40] applied to the discriminator. R1 penalty penalizes the discriminator for deviating from the Nash-equilibrium by penalizing the gradient on real data alone. The gradient penalty weight is set to 10. Adam [29] is utilized for optimization with β1 = 0 and β2 = 0.99. The learning rate is 0.0001 for both the generator and discriminator. All the models are trained using TPU for one million iterations on Image Net and 500,000 iterations on FFHQ and Celeb A-HQ. We set the mini-batch size to 256 for the image resolution of 128 128 and 256 256 while to 32 for the resolution of 1024 1024. To keep our setup simple, we do not employ any training tricks for GAN training, such as progressive growing, equalized learning rates, pixel normalization, noise injection, and mini-batch standard deviation that recent literature demonstrate to be cruicial for obtaining state-of-the-art results [25, 28]. When training on FFHQ and Celeb A-HQ, we utilize balanced consistency regularization (b CR) [72, 77] for additional regularization where images are augmented by flipping, color, translation, and cutout as in [76]. b CR enforces that for both real and generated images, two sets of augmentations applied to the same input image should yield the same output. b CR is only added to the discriminator with the weight of 10. We also decrease the learning rate by half to stabilize the training process when b CR is used. 4.3 Results on Image Net Unconditional generation. We start by evaluating the proposed Hi T on the Image Net 128 128 dataset, targeting the unconditional image generation setting for simplicity. In addition to recently reported state-of-the-art GANs, we implement a Conv Net-based generator following the widely-used architecture from [69] while using the exactly same training setup of the proposed Hi T (e.g., losses and R1 gradient penalty) denoted as Conv Net-R1 for a fair comparison. The results are shown in Table 1. We can see that Hi T outperforms the previous Conv Net-based methods by a notable margin in terms of both IS and FID scores. Note that as reported in [11], Big GAN [4] achieves 30.91 FID on unconditional Image Net, but its model is far larger (more than 70M parameters) than Hi T (around 30M parameters). Therefore, the results are not directly comparable. Even though, Hi T (30.83 FID) Table 1: Comparison with the state-of-the-art methods on the Image Net 128 128 dataset. is based on a supervised pre-trained Image Net classifier. Method FID IS Vanilla GAN [15] 54.17 14.01 Pac GAN2 [35] 57.51 13.50 MGAN [18] 58.88 13.22 Logo-GAN-AE [50] 50.90 14.44 Logo-GAN-RC [50] 38.41 18.86 SS-GAN (s BN) [9] 43.87 - Self-Conditioned GAN [36] 40.30 15.82 Conv Net-R1 37.18 19.55 Hi T (Ours) 30.83 21.64 Table 2: Reconstruction FID on the Image Net 256 256 dataset. We note that VQVAE-2 utilizes a hierarchical organization of VQ-VAE and thus has two codebooks Z. Method Embedding size and |Z| FID VQ-VAE [61] 32, 1024 75.19 DALL-E [47] 32, 8192 34.30 VQ-VAE-2 [48] 64, 512 32, 512 10.00 VQGAN [14] 16, 1024 8.00 VQ-Hi T (Ours) 16, 1024 6.37 Figure 3: Unconditional image generation results of Hi T trained on Image Net 128 128. is slightly better than Big GAN and achieves the state of the art among models with a similar capacity as shown in Table 1. We also note that [8, 36] leverage auxiliary regularization techniques and our method is complementary to them. Examples generated by Hi T on Image Net are shown in Figure 3, from which we observe nature visual details and diversity. Reconstruction. We are also interested in the reconstruction ability of the proposed Hi T and evaluate by employing it as a decoder for the vector quantised variational auto-encoder (VQ-VAE [61]), a state-of-the-art approach for visual representation learning. In addition to the reconstruction loss and adversarial loss, our Hi T-based VQ-VAE variant, namely VQ-Hi T, is also trained with the perceptual loss [24] following the setup of [14, 78]. Please refer to the supplementary materials for more details on the architecture design and model training. We evaluate the metric of reconstruction FID on Image Net 256 256 and report the results in Table 2. Our VQ-Hi T attains the best performance while providing significantly more compression (i.e., smaller embedding size and fewer number of codes in the codebook Z) than [47, 48, 61]. 4.4 Higher Resolution Generation Baselines. To demonstrate the utility of our approach for high-fidelity images, we compare to stateof-the-art techniques for generating images on Celeb A-HQ and FFHQ, focusing on two common resolutions of 256 256 and 1024 1024. The main competing method of the proposed Hi T is Style GAN [27, 28] a hypernetwork-based Conv Net archieving the best performance on these two datasets. On the FFHQ dataset, apart from our Conv Net-based counterparts, we also compare to the most recent INR-based methods including CIPS [1] and INR-GAN [54]. These two models are closely related to Hi T as the high-resolution stages of our approach can be viewed as a form of INR. Following the protocol of [1], we report the results of Style GAN2 [28] trained without style-mixing and path regularization of the generator on this dataset. Table 3: Comparison with the state-of-the-art methods on Celeb A-HQ (left) and FFHQ (right) with the resolutions of 256 256 and 1024 1024. b CR is not applied at the 1024 1024 resolution. FID (Celeb A-HQ) Method 256 1024 VAEBM [68] 20.38 - Style ALAE [45] 19.21 - PG-GAN [25] 8.03 - COCO-GAN [33] - 9.49 VQGAN [14] 10.70 - Style GAN [27] - 5.17 Hi T-B (Ours) 3.39 8.83 FID (FFHQ) Method 256 1024 U-Net GAN [52] 7.63 - Style ALAE [45] - 13.09 VQGAN [14] 11.40 - INR-GAN [54] 9.57 16.32 CIPS [1] 4.38 10.07 Style GAN2 [28] 3.83 4.41 Hi T-B (Ours) 2.95 6.37 Table 4: Comparison with the main competing methods in terms of number of network parameters, throughput, and FID on FFHQ 256 256. The throughput is measured on a single Tesla V100 GPU. Architecture Model #params (million) Throughput (images / sec) FID (FFHQ 256) Conv Net Style GAN2 [28] 30.03 95.79 3.83 INR CIPS [1] 45.90 27.27 4.38 INR-GAN [54] 107.03 266.45 9.57 Transformer Hi T-S 38.01 86.64 3.06 Hi T-B 46.22 52.09 2.95 Hi T-L 97.46 20.67 2.58 Results. We report the results in Table 3. Impressively, the proposed Hi T obtains the best FID scores at the resolution of 256 256 and sets the new state of the art on both datasets. Meanwhile, our performance is also competitive at the resolution of 1024 1024 but only slightly shy of Style GAN. This is due to our finding that conventional regularization techniques such as [77] cannot be directly applied to Transformer-based architectures for synthesizing ultra high-resolution images. We believe this triggers a new research direction and will explore it as our future work. It is also worth mentioning that our method consistently outperforms INR-based models, which suggests the importance of involving the self-attention mechanism into image generation. Table 4 provides a more detailed comparison to our main competing methods in terms of the number of parameters, throughput, and FID on FFHQ 256 256. We find that the proposed Hi T-S has a comparable runtime performance (throughput) with Style GAN while yielding better generative results. More importantly, FID scores can be further improved when larger variants of Hi T are employed. Figure 4 illustrates our synthetic face images on Celeb A-HQ. 4.5 Ablation Study We then evaluate the importance of different components of our model by its ablation on Image Net 128 128. A lighter-weight training configuration is employed in this study to reduce the training time. We start with a baseline architecture without any attention modules which reduces to a pure implicit neural function conditioned on the input latent code [3, 30]. We build upon this baseline by gradually incorporating cross-attention and self-attention modules. We compare the methods that perform attention without blocking (including standard self-attention [63], axial attention [17, 65]) and with blocking (including blocked local attention [62, 73]). We also compare a variant of our method where different types of attention are interleaved other than combined in attention heads. The results are reported in Table 5. As we can see, the latent code conditional INR baseline cannot fit training data favourably since it lacks an efficient way to capture the latent code during the generative process. Interestingly, incorporating the proposed cross-attention module improves the baseline and achieves the state of the art, thanks to its function of self-modulation. More importantly, we find that blocking is vital for attention: performing different types of attention after blocking can all improve the performance by a notable margin while the axial attention cannot. This is due to the fact that the Figure 4: Synthetic face images by Hi T-B trained on Celeb A-HQ 1024 1024 and 256 256. Table 5: Ablation study. We start with the INR-based generator [3, 30] conditioned on the input latent code and gradually improve it with the proposed attention components and their variations. O/M denotes out-of-memory error: the model cannot be trained for the batch size of one. Model configuration #params (million) Throughput (images / sec) FID IS Latent-code conditioned INR decoder [3, 30] 42.68 110.39 56.33 16.19 + Cross-attention for self-modulation 61.55 82.67 35.94 19.42 All-to-all self-attention [63] 67.60 - O/M O/M Axial attention [17, 20, 65] 67.60 74.21 35.15 19.79 Blocked local attention [62, 73] 67.60 75.54 33.70 19.96 Interleaving blocked regional and dilated attention 32.96 20.75 Multi-axis blocked self-attention (Ours) 32.23 20.96 + Balancing attention between axes (Full model) 67.60 75.33 31.87 21.32 image structure after blocking introduces a better inductive bias for images. At last, we observe that the proposed multi-axis blocked self-attention yields better FID scores than interleaving different attention outputs, and it achieves the best performance after balancing the attention axes. In Table 6, we investigate another form of our variations, where we incorporate attention in different number of stages across the generator. We can see that the more in the stack the attention is applied (low-resolution stages), the better the model performance will be. This study provides a validation for the effectiveness of self-attention operation in generative modeling. However, for feature resolutions larger than 64 64, we do not observe obvious performance gains brought by self-attention operations but a notable degradation in both training and inference time. 4.6 Effectiveness of Regularization We further explore the effectiveness of regularization for the proposed Transformer-based generator. Note that different from the recent interest in training GANs in the few-shot scenarios [26, 76], we study the influence of regularization for Conv Net-based and Transformer-based generators in the fulldata regime. We use the whole training set of FFHQ 256 256, and compare Hi T-based variants with Style GAN2 [26]. The regularization method is b CR [77]. As shown in Table 7, all variants of Hi T achieve much larger margins of improvement than Style GAN2. This confirms the finding [12, 23, 60] that Transformer-based architectures are much more data-hungry than Conv Nets in both classification and generative modeling settings, and thus strong data augmentation and regularization techniques are crucial for training Transformer-based generators. 5 Conclusion We present Hi T, a novel Transformer-based generator for high-resolution image generation based on GANs. To address the quadratic scaling problem of Transformers, we structure the low-resolution stages of Hi T following the design of Nested Transformer and enhance it by the proposed multi-axis blocked self-attention. In order to handle extremely long inputs in the high-resolution stages, we Table 6: Performance as a function of the number of self-attention stages on Image Net 128 128. The attention configuration is defined using the protocol [a, b], where a and b refer to the number of stages in the low-resolution and high-resolution stages of the model, respectively. Attention configuration [0, 5] [1, 4] [2, 3] [3, 2] [4, 1] #params (million) 61.55 66.01 67.19 67.52 67.60 Throughput (images / sec) 82.67 80.88 80.22 78.06 75.33 FID 35.94 34.16 33.69 32.72 31.87 Table 7: The effectiveness of b CR [77] on both Style GAN2 and Hi T-based variants. indicates the results of Style GAN2 are obtained from [26] which uses a lighter-weight configuration of [28]. + b CR [77] Style GAN2 [28] Hi T-S Hi T-B Hi T-L 5.28 6.07 5.30 5.13 3.91 3.06 2.95 2.58 FID 1.37 3.01 2.35 2.55 drop self-attention operations and reduce the model into implicit functions. We further improve the model performance by introducing a cross-attention module which plays a role of self-modulation. Our experiments demonstrate that Hi T achieves highly competitive performance for high-resolution image generation compared with its Conv Net-based counterparts. For future work, we will investigate Transformer-based architectures for discriminators in GANs and efficient regularization techniques for Transformer-based generators to synthesize ultra high-resolution images. Ethics statement. We note that although this paper does not uniquely raise any new ethical challenges, image generation is a field with several ethical concerns worth acknowledging. For example, there are known issues around bias and fairness, either in the representation of generated images [39] or the implicit encoding of stereotypes [55]. Additionally, such algorithms are vulnerable to malicious use [5], mainly through the development of deepfakes and other generated media meant to deceive the public [67], or as an image denoising technique that could be used for privacy-invading surveillance or monitoring purposes. We also note that the synthetic image generation techniques have the potential to mitigate bias and privacy issues for data collection and annotation. However, such techniques could be misused to produce misleading information, and researchers should explore the techniques responsibly. Acknowledgements We thank Chitwan Saharia, Jing Yu Koh, and Kevin Murphy for their feedback to the paper. We also thank Ashish Vaswani, Mohammad Taghi Saffar, and the Google Brain team for research discussions and technical assistance. This research has been partially funded by the following grants ARO MURI 805491, NSF IIS-1793883, NSF CNS-1747778, NSF IIS 1763523, DOD-ARO ACC-W911NF, NSF OIA-2040638 to Dimitris N. Metaxas. [1] Ivan Anokhin, Kirill Demochkin, Taras Khakhulin, Gleb Sterkin, Victor Lempitsky, and Denis Korzhenkov. Image generators with conditionally-independent pixel synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [2] Iz Beltagy, Matthew E Peters, and Arman Cohan. Longformer: The long-document transformer. ar Xiv preprint ar Xiv:2004.05150, 2020. [3] Tristan Bepler, Ellen D Zhong, Kotaro Kelley, Edward Brignole, and Bonnie Berger. Explicitly disentangling image content from translation and rotation with spatial-VAE. In Advances in Neural Information Processing Systems (Neur IPS), pages 15409 15419, 2019. [4] Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale GAN training for high fidelity natural image synthesis. In Proceedings of the International Conference on Learning Representations (ICLR), 2019. [5] Miles Brundage, Shahar Avin, Jack Clark, Helen Toner, Peter Eckersley, Ben Garfinkel, Allan Dafoe, Paul Scharre, Thomas Zeitzoff, Bobby Filar, et al. The malicious use of artificial intelligence: Forecasting, prevention, and mitigation. ar Xiv preprint ar Xiv:1802.07228, 2018. [6] Jiezhang Cao, Yawei Li, Kai Zhang, and Luc Van Gool. Video Super-Resolution Transformer. ar Xiv preprint ar Xiv:2106.06847, 2021. [7] Chun-Fu Chen, Quanfu Fan, and Rameswar Panda. Cross Vi T: Cross-attention multi-scale vision transformer for image classification. ar Xiv preprint ar Xiv:2103.14899, 2021. [8] Ting Chen, Mario Lucic, Neil Houlsby, and Sylvain Gelly. On self modulation for generative adversarial networks. In Proceedings of the International Conference on Learning Representations (ICLR), 2019. [9] Ting Chen, Xiaohua Zhai, Marvin Ritter, Mario Lucic, and Neil Houlsby. Self-supervised GANs via auxiliary rotation loss. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 12154 12163, 2019. [10] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), pages 215 223, 2011. [11] Jeff Donahue and Karen Simonyan. Large scale adversarial representation learning. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [12] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations (ICLR), 2021. [13] Emilien Dupont, Yee Whye Teh, and Arnaud Doucet. Generative models as distributions of functions. ar Xiv preprint ar Xiv:2102.04776, 2021. [14] Patrick Esser, Robin Rombach, and Björn Ommer. Taming transformers for high-resolution image synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [15] Ian J Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems (Neur IPS), 2014. [16] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems (Neur IPS), pages 6629 6640, 2017. [17] Jonathan Ho, Nal Kalchbrenner, Dirk Weissenborn, and Tim Salimans. Axial attention in multidimensional transformers. ar Xiv preprint ar Xiv:1912.12180, 2019. [18] Quan Hoang, Tu Dinh Nguyen, Trung Le, and Dinh Phung. MGAN: Training generative adversarial nets with multiple generators. In Proceedings of the International Conference on Learning Representations (ICLR), 2018. [19] Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 1501 1510, 2017. [20] Zilong Huang, Xinggang Wang, Lichao Huang, Chang Huang, Yunchao Wei, and Wenyu Liu. CCNet: Criss-cross attention for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 603 612, 2019. [21] Drew A Hudson and C Lawrence Zitnick. Generative adversarial transformers. In Proceedings of the International Conference on Machine Learning (ICML), 2021. [22] Andrew Jaegle, Felix Gimeno, Andrew Brock, Andrew Zisserman, Oriol Vinyals, and Joao Carreira. Perceiver: General perception with iterative attention. ar Xiv preprint ar Xiv:2103.03206, 2021. [23] Yifan Jiang, Shiyu Chang, and Zhangyang Wang. Trans GAN: Two transformers can make one strong GAN. ar Xiv preprint ar Xiv:2102.07074, 2021. [24] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and superresolution. In Proceedings of the European Conference on Computer Vision (ECCV), pages 694 711, 2016. [25] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs for improved quality, stability, and variation. In Proceedings of the International Conference on Learning Representations (ICLR), 2018. [26] Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Training generative adversarial networks with limited data. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [27] Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4401 4410, 2019. [28] Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Analyzing and improving the image quality of Style GAN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 8110 8119, 2020. [29] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), 2014. [30] Marian Kleineberg, Matthias Fey, and Frank Weichert. Adversarial generation of continuous implicit shape representations. In Eurographics, 2020. [31] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images, 2009. [32] Kwonjoon Lee, Huiwen Chang, Lu Jiang, Han Zhang, Zhuowen Tu, and Ce Liu. Vi TGAN: Training GANs with Vision Transformers. ar Xiv preprint ar Xiv:2107.04589, 2021. [33] Chieh Hubert Lin, Chia-Che Chang, Yu-Sheng Chen, Da-Cheng Juan, Wei Wei, and Hwann-Tzong Chen. COCO-GAN: Generation by Parts via Conditional Coordinating. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 4512 4521, 2019. [34] Chieh Hubert Lin, Hsin-Ying Lee, Yen-Chi Cheng, Sergey Tulyakov, and Ming-Hsuan Yang. Infinity GAN: Towards infinite-resolution image synthesis. ar Xiv preprint ar Xiv:2104.03963, 2021. [35] Zinan Lin, Ashish Khetan, Giulia Fanti, and Sewoong Oh. Pac GAN: The power of two samples in generative adversarial networks. In Advances in Neural Information Processing Systems (Neur IPS), 2018. [36] Steven Liu, Tongzhou Wang, David Bau, Jun-Yan Zhu, and Antonio Torralba. Diverse image generation via self-conditioned GANs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 14286 14295, 2020. [37] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. ar Xiv preprint ar Xiv:2103.14030, 2021. [38] Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2015. [39] Sachit Menon, Alexandru Damian, Shijia Hu, Nikhil Ravi, and Cynthia Rudin. PULSE: Self-supervised photo upsampling via latent space exploration of generative models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2437 2445, 2020. [40] Lars Mescheder, Andreas Geiger, and Sebastian Nowozin. Which training methods for GANs do actually converge? In Proceedings of the International Conference on Machine Learning (ICML), pages 3481 3490, 2018. [41] Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3D reconstruction in function space. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4460 4470, 2019. [42] Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Ne RF: Representing scenes as neural radiance fields for view synthesis. In Proceedings of the European Conference on Computer Vision (ECCV), pages 405 421, 2020. [43] Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deep SDF: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 165 174, 2019. [44] Niki Parmar, Ashish Vaswani, Jakob Uszkoreit, Lukasz Kaiser, Noam Shazeer, Alexander Ku, and Dustin Tran. Image transformer. In Proceedings of the International Conference on Machine Learning (ICML), pages 4055 4064, 2018. [45] Stanislav Pidhorskyi, Donald A Adjeroh, and Gianfranco Doretto. Adversarial latent autoencoders. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 14104 14113, 2020. [46] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. ar Xiv preprint ar Xiv:2103.00020, 2021. [47] Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. Zero-shot text-to-image generation. ar Xiv preprint ar Xiv:2102.12092, 2021. [48] Ali Razavi, Aaron van den Oord, and Oriol Vinyals. Generating diverse high-fidelity images with VQVAE-2. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [49] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Image Net large scale visual recognition challenge. International Journal of Computer Vision, 115(3):211 252, 2015. [50] Alexander Sage, Eirikur Agustsson, Radu Timofte, and Luc Van Gool. Logo synthesis and manipulation with clustered generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5879 5888, 2018. [51] Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, Xi Chen, and Xi Chen. Improved techniques for training GANs. In Advances in Neural Information Processing Systems (Neur IPS), 2016. [52] Edgar Schonfeld, Bernt Schiele, and Anna Khoreva. A U-Net Based Discriminator for Generative Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 8207 8216, 2020. [53] Wenzhe Shi, Jose Caballero, Ferenc Huszár, Johannes Totz, Andrew P Aitken, Rob Bishop, Daniel Rueckert, and Zehan Wang. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1874 1883, 2016. [54] Ivan Skorokhodov, Savva Ignatyev, and Mohamed Elhoseiny. Adversarial generation of continuous images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [55] Ryan Steed and Aylin Caliskan. Image representations learned with unsupervised pre-training contain human-like biases. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pages 701 713, 2021. [56] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2818 2826, 2016. [57] Yu Tian, Xi Peng, Long Zhao, Shaoting Zhang, and Dimitris N Metaxas. CR-GAN: Learning complete representations for multi-view generation. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), pages 942 948, 2018. [58] Ilya Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Daniel Keysers, Jakob Uszkoreit, Mario Lucic, et al. MLP-Mixer: An all-MLP Architecture for Vision. ar Xiv preprint ar Xiv:2105.01601, 2021. [59] Hugo Touvron, Piotr Bojanowski, Mathilde Caron, Matthieu Cord, Alaaeldin El-Nouby, Edouard Grave, Armand Joulin, Gabriel Synnaeve, Jakob Verbeek, and Hervé Jégou. Res MLP: Feedforward networks for image classification with data-efficient training. ar Xiv preprint ar Xiv:2105.03404, 2021. [60] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. ar Xiv preprint ar Xiv:2012.12877, 2020. [61] Aaron van den Oord, Oriol Vinyals, and Koray Kavukcuoglu. Neural discrete representation learning. In Advances in Neural Information Processing Systems (Neur IPS), pages 6309 6318, 2017. [62] Ashish Vaswani, Prajit Ramachandran, Aravind Srinivas, Niki Parmar, Blake Hechtman, and Jonathon Shlens. Scaling local self-attention for parameter efficient visual backbones. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [63] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems (Neur IPS), pages 6000 6010, 2017. [64] Huiyu Wang, Yukun Zhu, Hartwig Adam, Alan Yuille, and Liang-Chieh Chen. Ma X-Deep Lab: End-to-end panoptic segmentation with mask transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [65] Huiyu Wang, Yukun Zhu, Bradley Green, Hartwig Adam, Alan Yuille, and Liang-Chieh Chen. Axial Deep Lab: Stand-alone axial-attention for panoptic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 108 126, 2020. [66] Xiaolong Wang, Ross Girshick, Abhinav Gupta, and Kaiming He. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 7794 7803, 2018. [67] Mika Westerlund. The emergence of deepfake technology: A review. Technology Innovation Management Review, 9(11), 2019. [68] Zhisheng Xiao, Karsten Kreis, Jan Kautz, and Arash Vahdat. VAEBM: A Symbiosis between Variational Autoencoders and Energy-based Models. In Proceedings of the International Conference on Learning Representations (ICLR), 2021. [69] Han Zhang, Ian Goodfellow, Dimitris Metaxas, and Augustus Odena. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning (ICML), pages 7354 7363, 2019. [70] Han Zhang, Jing Yu Koh, Jason Baldridge, Honglak Lee, and Yinfei Yang. Cross-modal contrastive learning for text-to-image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. [71] Han Zhang, Tao Xu, Hongsheng Li, Shaoting Zhang, Xiaogang Wang, Xiaolei Huang, and Dimitris N Metaxas. Stack GAN: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 5907 5915, 2017. [72] Han Zhang, Zizhao Zhang, Augustus Odena, and Honglak Lee. Consistency regularization for generative adversarial networks. In Proceedings of the International Conference on Learning Representations (ICLR), 2020. [73] Zizhao Zhang, Han Zhang, Long Zhao, Ting Chen, and Tomas Pfister. Aggregating nested transformers. ar Xiv preprint ar Xiv:2105.12723, 2021. [74] Long Zhao, Xi Peng, Yu Tian, Mubbasir Kapadia, and Dimitris Metaxas. Learning to forecast and refine residual motion for image-to-video generation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 387 403, 2018. [75] Long Zhao, Xi Peng, Yu Tian, Mubbasir Kapadia, and Dimitris N Metaxas. Towards image-to-video translation: A structure-aware approach via multi-stage generative adversarial networks. International Journal of Computer Vision, 128(10):2514 2533, 2020. [76] Shengyu Zhao, Zhijian Liu, Ji Lin, Jun-Yan Zhu, and Song Han. Differentiable augmentation for dataefficient gan training. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [77] Zhengli Zhao, Sameer Singh, Honglak Lee, Zizhao Zhang, Augustus Odena, and Han Zhang. Improved consistency regularization for GANs. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 2021. [78] Jiapeng Zhu, Yujun Shen, Deli Zhao, and Bolei Zhou. In-domain GAN inversion for real image editing. In Proceedings of the European Conference on Computer Vision (ECCV), pages 592 608, 2020. [79] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2223 2232, 2017.