# projected_gans_converge_faster__245ed88d.pdf Projected GANs Converge Faster Axel Sauer1,2 Kashyap Chitta1,2 Jens Müller3 Andreas Geiger1,2 1University of Tübingen 2Max Planck Institute for Intelligent Systems, Tübingen 3Computer Vision and Learning Lab, University Heidelberg 2{firstname.lastname}@tue.mpg.de 3{firstname.lastname}@iwr.uni-heidelberg.de Generative Adversarial Networks (GANs) produce high-quality images but are challenging to train. They need careful regularization, vast amounts of compute, and expensive hyper-parameter sweeps. We make significant headway on these issues by projecting generated and real samples into a fixed, pretrained feature space. Motivated by the finding that the discriminator cannot fully exploit features from deeper layers of the pretrained model, we propose a more effective strategy that mixes features across channels and resolutions. Our Projected GAN improves image quality, sample efficiency, and convergence speed. It is further compatible with resolutions of up to one Megapixel and advances the state-of-the-art Fréchet Inception Distance (FID) on twenty-two benchmark datasets. Importantly, Projected GANs match the previously lowest FIDs up to 40 times faster, cutting the wall-clock time from 5 days to less than 3 hours given the same computational resources. 1.5 5 10 15 20 25 Training Time [hours] FID (log scale) Style GAN2-ADA Style GAN2-ADA (converged) Projected GAN Projected GAN (converged) Style GAN2-ADA Projected GAN Figure 1: Convergence with Projected GANs. Evolution of samples for a fixed latent code during training on the AFHQ-Dog dataset [5]. We find that discriminating features in the projected feature space speeds up convergence and yields lower FIDs. This finding is consistent across many datasets. 1 Introduction A Generative Adversarial Network (GAN) consists of a generator and a discriminator. For image synthesis, the generator s task is to generate an RGB image; the discriminator aims to distinguish real from fake samples. On closer inspection, the discriminator s task is two-fold: First, it projects the real and fake samples into a meaningful space, i.e., it learns a representation of the input space. Second, it discriminates based on this representation. Unfortunately, training the discriminator jointly with the generator is a notoriously hard task. While discriminator regularization techniques help to balance the adversarial game [31], standard regularization methods like gradient penalties [36] are susceptible to hyperparameter choices [26] and can lead to a substantial decrease in performance [4]. In this paper, we explore the utility of pretrained representations to improve and stabilize GAN training. Using pretrained representations has become ubiquitous in computer vision [29, 30, 48] 35th Conference on Neural Information Processing Systems (Neur IPS 2021). and natural language processing [18,45,47]. While combining pretrained perceptual networks [58] with GANs for image-to-image translation has led to impressive results [14,49,59,64], this idea has not yet materialized for unconditional noise-to-image synthesis. Indeed, we confirm that a naïve application of this idea does not lead to state-of-the-art results (Section 4) as strong pretrained features enable the discriminator to dominate the two-player game, resulting in vanishing gradients for the generator [2]. In this work, we demonstrate how these challenges can be overcome and identify two key components for exploiting the full potential of pretrained perceptual feature spaces for GAN training: feature pyramids to enable multi-scale feedback with multiple discriminators and random projections to better utilize deeper layers of the pretrained network. We conduct extensive experiments on small and large datasets with a resolution of up to 10242 pixels. Across all datasets, we demonstrate state-of-the-art image synthesis results at significantly reduced training time (Fig. 1). We also find that Projected GANs increase data efficiency and avoid the need for additional regularization, rendering expensive hyperparameter sweeps unnecessary. Code, models, and supplementary videos can be found on the project page https://sites.google.com/view/ projected-gan. 2 Related Work We categorize related work into two main areas: pretraining for GANs and discriminator design. Pretrained Models for GAN Training. Work on leveraging pretrained representations for GANs can be divided into two categories: First, transferring parts of a GAN to a new dataset [15,38,65,71] and, second, using pretrained models to control and improve GANs. The latter is advantageous as pretraining does not need to be adversarial. Our work falls into this second category. Pretrained models can be used as a guiding mechanism to disentangle causal generative factors [54], for text-driven image manipulation [44], matching the generator activations to inverted classifiers [19,56], or to generate images via gradient ascent in the latent space of a generator [41]. The non-adversarial approach of [53] learns generative models with moment matching in pretrained models; however, the results remain far from competitive to standard GANs. An established method is the combination of adversarial and perceptual losses [21]. Commonly, the losses are combined additively [10,14,32,52,64]. Additive combination, however, is only possible if a reconstruction target is available, e.g., in paired image-toimage translation settings [74]. Instead of providing the pretrained network with a reconstruction target, Sungatullina et al. [59] propose to optimize an adversarial loss on frozen VGG features [58]. They show that their approach improves Cycle GAN [74] on image translation tasks. In a similar vein, [49] recently proposed a different perceptual discriminator. They utilize a pretrained VGG and connect its features with the prediction of a pretrained segmentation network. The combined features are fed into multiple discriminators at different scales. The two last approaches are specific to the image-toimage translation task. We demonstrate that these methods do not work well for the more challenging unconditional setting where the entire image content is synthesized from a random latent code. Discriminator Design. Much work on GANs focuses on novel generator architectures [4,26,27,69], while the discriminator often remains close to a vanilla convolutional neural network or mirrors the generator. Notable exceptions are [55,70] which utilize an encoder-decoder discriminator architecture. However, in contrast to us, they neither use pretrained features nor random projections. A different line of work considers a setup with multiple discriminators, applied to either the generated RGB image [8, 13] or low-dimensional projections thereof [1, 40]. The use of several discriminators promises improved sample diversity, training speed, and training stability. However, these approaches are not utilized in current state-of-the-art systems because of diminishing returns compared to the increased computational effort. Providing multi-scale feedback with one or multiple discriminators has been helpful for both image synthesis [23,24] and image-to-image translation [43,64]. While these works interpolate the RGB image at different resolutions, our findings indicate the importance of multi-scale feature maps, showing parallels to the success of pyramid networks for object detection [34]. Lastly, to prevent overfitting of the discriminator, differentiable augmentation methods have recently been proposed [25,63,72,73]. We find that adopting these strategies helps exploit the full potential of pretrained representations for GAN training. 3 Projected GANs GANs aim to model the distribution of a given training dataset. A generator G maps latent vectors z sampled from a simple distribution Pz (typically a normal distribution) to corresponding generated samples G(z). The discriminator D then aims to distinguish real samples x Px from the generated samples G(z) PG(z). This basic idea results in the following minimax objective min G max D Ex[log D(x)] + Ez[log(1 D(G(z)))] (1) We introduce a set of feature projectors {Pl} which map real and generated images to the discriminator s input space. Projected GAN training can thus be formulated as follows min G max {Dl} Ex[log Dl(Pl(x))] + Ez[log(1 Dl(Pl(G(z)))))] (2) where {Dl} is a set of independent discriminators operating on different feature projections. Note that we keep {Pl} fixed in (2) and only optimize the parameters of G and {Dl}. The feature projectors {Pl} should satisfy two necessary conditions: they should be differentiable and provide sufficient statistics of their inputs, i.e., they should preserve important information. Moreover, we aim to find feature projectors {Pl} which turn the (difficult to optimize) objective in (1) into an objective more amenable to gradient-based optimization. We now show that a projected GAN indeed matches the distribution in the projected feature space, before specifying the details of our feature projectors. 3.1 Consistency The projected GAN objective in (2) no longer optimizes directly to match the true distribution PT . To understand the training properties under ideal conditions, we consider a more generalized form of the consistency theorem of [40]: Theorem 1. Let PT denote the density of the true data distribution and PG the density of the distribution the Generator G produces. Let Pl T and Pl G be the functional composition of the differentiable and fixed function Pl and the true/generated data distribution, and y be the transformed input to the discriminator. For a fixed G, the optimal discriminators are given by D l,G(y) = PPl T (y) PPl T (y) + PPl G(y) for all l L. In this case, the optimal G under (2) is achieved iff PPl T = PPl G for all l L. A proof of the theorem is provided in the appendix. From the theorem, we conclude that a feature projector Pl with its associated discriminator Dl encourages the generator to match the true distribution along the marginal through Pl. Therefore, at convergence, G matches the generated and true distributions in feature space. The theorem also holds when using stochastic data augmentations [25] before the deterministic projections Pl. 3.2 Model Overview Projecting to and training in pretrained feature spaces opens up a realm of new questions which we address below. This section will provide an overview of the general system and is followed by extensive ablations of each design choice. As our feature projections affect the discriminator, we focus on Pl and Dl in this section and postpone the discussion of generator architectures to Section 5. Multi-Scale Discriminators. We obtain features from four layers Ll of a pretrained feature network F at resolutions (L1 = 642, L2 = 322, L3 = 162, L4 = 82). We associate a separate discriminator Dl with the features at layer Ll, respectively. Each discriminator Dl uses a simple convolutional architecture with spectral normalization [37] at each convolutional layer. We observe better performance if all discriminators output logits at the same resolution (42). Accordingly, we use fewer down-sampling blocks for lower resolution inputs. Following common practice, we sum all logits for computing the overall loss. For the generator pass, we sum the losses of all discriminators. More complex strategies [1,13] did not improve performance in our experiments. Random Projections. We observe that features at deeper layers are significantly harder to cover, as evidenced by our experiments in Section 4. We hypothesize that a discriminator can focus on a subset of the feature space while wholly disregarding other parts. This problem might be especially prominent in the deeper, more semantic layers. Therefore, we propose two different strategies to dilute prominent features, encouraging the discriminator to utilize all available information equally. Common to both strategies is that they mix features using differentiable random projections which are fixed, i.e., after random initialization, the parameters of these layers are not trained. Real/Generated Image Feature Network Figure 2: CCM (dashed blue arrows) employs 1 1 convolutions with random weights. Cross-Channel Mixing (CCM). Empirically, we found two properties to be desirable: (i) the random projection should be information preserving to leverage the full representational power of F, and (ii) it should not be trivially invertible. The easiest way to mix across channels is a 1 1 convolution. A 1 1 convolution with an equal number of output and input channels is a generalization of a permutation [28] and consequently preserves information about its input. In practice, we find that more output channels lead to better performance as the mapping remains injective and therefore information preserving. Kingma et al. [28] initialize their convolutional layers as a random rotation matrix as a good starting point for optimization. We do not find this to improve GAN performance (see Appendix), arguably since it violates (ii). We therefore randomly initialize the weights of the convolutional layer via Kaiming initialization [16]. Note that we do not add any activation functions. We apply this random projection at each of the four scales and feed the transformed feature to the discriminator as depicted in Fig. 2. Real/Generated Image Feature Network Figure 3: CSM (dashed red arrows) adds random 3 3 convolutions and bilinear upsampling, yielding a U-Network. Cross-Scale Mixing (CSM). To encourage feature mixing across scales, CSM extends CCM with random 3 3 convolutions and bilinear upsampling, yielding a U-Net [50] architecture, see Fig. 3. However, our CSM block is simpler than a vanilla U-Net [50]: we only use a single convolutional layer at each scale. As for CCM, we utilize Kaiming initialization for all weights. Pretrained Feature Networks. We ablate over varying feature networks. First, we investigate different versions of Efficient Nets, which allow for direct control over model size versus performance. Efficient Nets are image classification models trained on Image Net [7] and designed to provide favorable accuracy-compute tradeoffs. Second, we use Res Nets of varying sizes. To analyze the dependency on Image Net features (Section 4.3), we also consider R50-CLIP [46], a Res Net optimized with a contrastive language-image objective on a dataset of 400 million (image, text) pairs. Lastly, we utilize a vision transformer architecture (Vi TBase) [9] and its efficient follow-up (Dei T-small distilled) [62]. We do not choose an inception network [60] to avoid strong correlations with the evaluation metric FID [17]. In the appendix, we also evaluate several other neural and non-neural metrics to rule out correlations. These additional metrics reflect the rankings obtained by FID. In the following, we conduct a systematic ablation study to analyze the importance and best configuration of each component in our Projected GAN model, before comparing it to the state-of-the-art. 4 Ablation Study To determine the best configuration of discriminators, mixing strategy, and pretrained feature network, we conduct experiments on LSUN-Church [67], which is medium-sized (126k images) and reasonably visually complex, using a resolution of 2562 pixels. For the generator G we use the generator architecture of Fast GAN [35], consisting of several upsampling blocks, with additional skip-layerexcitation blocks. Using a hinge loss [33], we train with a batch size of 64 until 1 million real images have been shown to the discriminator, a sufficient amount for G to reach values close to convergence. If not specified otherwise, we use an Efficient Net-Lite1 [61] feature network in this section. We found that discriminator augmentation [25,63,72,73] consistently improves the performance of all methods, and is required to reach state-of-the-art performance. We leverage differentiable data-augmentation [72] which we found to yield the best results in combination with the Fast GAN generator. Discriminator(s) rel-FD1 rel-FD2 rel-FD3 rel-FD4 rel FID No Projection on L1 0.56 0.32 0.31 0.55 0.66 on L1, L2 0.35 0.21 0.23 0.47 0.53 on L1, L2, L3 0.42 0.26 0.28 0.64 0.90 on L1, L2, L3, L4 0.46 0.34 0.38 0.79 1.15 on L2, L3, L4 0.95 0.67 0.71 1.19 1.99 on L3, L4 2.14 1.41 1.18 1.99 3.46 on L4 10.92 5.74 2.56 2.79 5.08 Perceptual D 2.98 1.76 1.20 1.89 2.73 on L1 0.27 0.21 0.26 0.50 0.59 on L1, L2 0.27 0.18 0.21 0.41 0.48 on L1, L2, L3 0.31 0.25 0.24 0.54 0.67 on L1, L2, L3, L4 0.53 0.34 0.34 0.59 0.77 Perceptual D 5.33 3.06 2.14 1.09 4.77 on L1 0.34 0.25 0.19 0.35 0.44 on L1, L2 0.21 0.18 0.16 0.27 0.31 on L1, L2, L3 0.41 0.26 0.17 0.23 0.29 on L1, L2, L3, L4 0.26 0.16 0.13 0.16 0.24 Perceptual D 2.53 1.37 0.89 0.43 2.13 Table 1: Feature Space Fréchet Distances. We aim to find the best combination of discriminators and random projections to fit the distributions in feature network F. We show the relative FD at different layers of F (rel-FDi) between 50k generated and real images on LSUN-Church. rel-FDi is normalized using the baseline Fréchet Distances for a model with a standard single RGB image discriminator. Hence, values > 1 indicate worse performance than the RGB baseline. We report rel-FD for four layers of an Efficient Net (L1, L2, L3 and L4 from shallow to deep), as well as relative Fréchet Inception Distance (FID) [17]. Note that rel-FDi should not be compared between different feature spaces, i.e., only within-column comparisons are meaningful. Blue boxes highlight the layers which we supervise via independent discriminators. The green box corresponds to a perceptual discriminator [59], which takes in all feature maps at once. 4.1 Which feature network layers are most informative? We first investigate the relevance of independent multi-scale discriminators. For this experiment, we do not use feature mixing. To measure how well G fits a particular feature space, we employ the Fréchet Distance (FD) [12] on the spatially pooled features denoted as FDi for layer i. FDs across different feature spaces are not directly comparable. Therefore, we train a GAN baseline with a standard RGB discriminator, record FDRGB i at each layer and quantify the relative improvement via the fraction rel-FDi = FDi/FDRGB i . We also investigate a perceptual discriminator [59], where feature maps are fed into different layers of the same discriminator to predict a single logit. The results in Table 1 (No Projection) show that two discriminators are better than one and improve over the vanilla RGB baseline. Surprisingly, adding discriminators at deep layers hurts performance. We conclude that these more semantic features do not respond well to direct adversarial losses. We also experimented with discriminators at resized versions of the original image, but could not find a setting of hyperparameters and architectures that improves over the single image baseline. Omitting the discriminators on the shallow features decreases performance, which is anticipated, as these layers contain most of the information about the original image. A similar effect has been observed for feature inversion [11] the deeper the layer, the harder it is to reconstruct its input. Lastly, we observe that independent discriminators outperform the perceptual discriminator by a significant margin. Efficient Net Res Net Transformer lite0 lite1 lite2 lite3 lite4 R18 R50 R50-CLIP Dei T Vi T Params (M) 2.96 3.72 4.36 6.42 11.15 11.18 23.51 23.53 92.36 317.52 IN top-1 75.48 76.64 77.47 79.82 81.54 69.75 79.04 N/A 85.42 85.16 FID 2.53 1.65 1.69 1.79 2.35 4.16 4.40 3.80 2.46 12.38 Table 2: Pretrained Feature Networks Study. We train the projected GAN with different pretrained feature networks. We find that compact Efficient Nets outperform both Res Nets and Transformers. 4.2 How can we best utilize the pretrained features? Given the insights from the previous section, we aim to improve the utilization of deep features. For this experiment, we only investigate configurations that include discriminators at high resolutions. Table 1 (CCM and CCM + CSM) presents the results for both mixing strategies. CCM moderately decreases the FDs across all settings, confirming our hypothesis that mixing channels results in better feedback for the generator. When adding CSM, we achieve another notable improvement across all configurations. Especially rel-FDi at deeper layers are significantly decreased, demonstrating CSM s usefulness to leverage deep semantic features. Interestingly, we observe that the best performance is now obtained by combining all four discriminators. A perceptual discriminator is again inferior to multiple discriminators. We remark that integrating the original image, via an independent discriminator or CCM or CSM always resulted in worse performance. This failure suggests that naïvely combining non-projected with projected adversarial optimization impairs training dynamics. 4.3 Which feature network architecture is most effective? Using the best setting determined by the experiments above (CCM + CSM with four discriminators), we study the effectiveness of various perceptual feature network architectures for Projected GAN training. To ensure convergence, also for larger architectures, we train for 10 million images. Table 2 reports the FIDs achieved on LSUN-Church. Surprisingly, we find that there is no correlation with Image Net accuracy. On the contrary, we observe lower FIDs for smaller models (e.g., Efficient Netslite). This observation indicates that a more compact representation is beneficial while at the same time reducing computational overhead and consequently training time. R50-CLIP slightly outperforms its R50 counterpart, indicating that Image Net features are not required to achieve low FID. For the sake of completeness, we also train with randomly initialized feature networks, which, however, converge to much higher FID values (see Appendix). In the following, we thus use Efficient Net-Lite1 as our feature network. 5 Comparison to State-of-the-Art This section conducts a comprehensive analysis demonstrating the advantages of Projected GANs with respect to state-of-the-art models. Our experiments are structured into three sections: evaluation of convergence speed and data efficiency (5.1), and comparisons on large (5.2) and small (5.3) benchmark datasets. We cover a wide variety of datasets in terms of size (hundreds to millions of samples), resolution (2562 to 10242), and visual complexity (clip-art, paintings, and photographs). Evaluation Protocol. We measure image quality using the Fréchet Inception Distance (FID) [17]. Following [26, 27], we report the FID between 50k generated and all real images. We select the snapshot with the best FID for each method. In addition to image quality, we include a metric to evaluate convergence. As in [25], we measure training progress based on the number of real images shown to the discriminator (Imgs). We report the number of images required by the model for the FID to reach values within 5% of the best FID over training. In the appendix, we also report other metrics that are less benchmarked in GAN literature: KID [3], Sw AV-FID [39], precision and recall [51]. Unless otherwise specified, we follow the evaluation protocol of [20] to facilitate fair comparisons. Specifically, we compare all approaches given the same fixed number of images (10 million). With this setting, each experiment takes roughly 100-200 GPU hours on a NVIDIA V100, for more details we refer to the appendix. Baselines. We use Style GAN2-ADA [25] and Fast GAN [35] as baselines. Style GAN2-ADA is the strongest model on most datasets in terms of sample quality, whereas Fast GAN excels in training speed. We implement these baselines and our Projected GANs within the codebase provided by the authors of Style GAN2-ADA [25]. For each model, we ran two kinds of data augmentation: 2 4 6 8 10 12 14 Training Progress [# million real images] FID (log scale) Style GAN2-ADA Projected Style GAN2-ADA Fast GAN Projected Fast GAN 1 10 70 Dataset Size [k Images] FID (log scale) Style GAN2-ADA Fast GAN Projected Fast GAN Figure 4: Training Properties. Left: Projected Fast GAN surpasses the best FID of Style GAN2 (at 88 M images) after just 1.1 M images on LSUN-Church. Right: Projected Fast GAN yields significantly improved FID scores, even when using subsets of CLEVR with 1k and 10k samples. 100k 200k 300k 400k 500k 600k 700k 800k 900k 1M Figure 5: Training progress on LSUN church at 2562 pixels. Shown are samples for a fixed noise vector z over k images. From top to bottom: Fast GAN, Style GAN2-ADA, Projected GAN. differentiable data-augmentation [72] and adaptive discriminator augmentation [25]. We select the better performing augmentation strategy per model. For all baselines and datasets, we perform data amplification through x-flips. Projected GANs use the same generator and discriminator architecture and training hyperparameters (learning rate and batch size) for all experiments. For high-resolution image generation, additional upsampling blocks are included in the generator to match the desired output resolution. We carefully tune all hyper-parameters for both baselines for best results: we find that Fast GAN is sensitive to the choice of batch size, and Style GAN2-ADA to the learning rate and R1 penalty. The appendix documents additional implementation details used in each of our experiments. 5.1 Convergence Speed and Data Efficiency Following [20] and [68], we analyze the training properties of Projected GANs on LSUN-Church at an image resolution of 2562 pixels and on the 70k CLEVR dataset [22]. In this section, we also train longer than 10 M images if necessary, as we are interested in convergence properties. Convergence Speed. We apply projected GAN training for both the style-based generator of Style GAN2 and the standard generator with a single input noise vector of Fast GAN. As shown in Fig. 4 (left), Fast GAN converges quickly but saturates at a high FID. Style GAN2 converges more slowly (88 M images) but reaches a lower FID. Projected GAN training improves both generators. Particularly for Fast GAN, improvements in both convergence speed and final FID are significant while improvements for Style GAN2 are less pronounced. Remarkably, Projected Fast GAN reaches the previously best FID of Style GAN2 after experiencing only 1.1 M images as compared to 88 M of Style GAN2. In wall clock time, this corresponds to less than 3 hours instead of 5 days. Hence, from now on, we utilize the Fast GAN generator and refer to this model simply as Projected GAN. Fig. 5 shows samples for a fixed noise vector z during training on LSUN-Church. For both Fast GAN and Style GAN, patches of texture gradually morph into a global structure. For Projected GAN, we directly observe the emergence of structure which becomes more detailed over time. Interestingly, the Projected GAN latent space appears to be very volatile, i.e., for fixed z the images undergo significant perceptual changes during training. In the non-projected cases, these changes are more gradual. We hypothesize that this induced volatility might be due to the discriminator providing more semantic feedback compared to conventional RGB losses. Such semantic feedback could introduce more stochasticity during training which in turn improves convergence and performance. We also observed that the signed real logits of the discriminator remain at the same level throughout training (see Appendix). Stable signed logits indicate that the discriminator does not suffer from overfitting. Sample Efficiency. The use of pretrained models is generally linked to improved sample efficiency. To evaluate this property, we also created two subsets of the 70k CLEVR dataset by randomly subsampling 10k and 1k images from it, respectively. As depicted in Fig. 4 (right), our Projected GAN significantly improves over both baselines across all dataset splits. 5.2 Large Datasets Besides CLEVR and LSUN-Church, we benchmark Projected GANs against various state-of-the-art models on three other large datasets: LSUN-Bedroom [67] (3M indoor bedroom scenes), FFHQ [26] (70k images of faces) and Cityscapes [6] (25k driving scenes captured from a vehicle). For all datasets, we use an image resolution of 2562 pixels. As Cityscapes and CLEVR images are not of aspect ratio 1:1 we resize them to 2562 for training. Besides Style GAN2-ADA and Fast GAN, we compare against SAGAN [69] and GANsformers [20]. All models were trained for 10 M images. For the large datasets, we also report numbers for Style GAN2 trained for more than 10 M images to report the lowest FID values achieved in previous literature (denoted as Style GAN2*). In the appendix, we report results on nine more large datasets. Table 3 shows that the Projected GAN outperforms all state-of-the-art models in terms of FID values on all datasets by a large margin. For example, on LSUN-Bedroom, it achieves an FID value of 1.52 compared to 6.15 by GANsformer, the previously best model in this setting. Projected GAN achieves state-of-the-art FID values remarkably fast, e.g., on LSUN-church, it achieves an FID value of 3.18 after 1.1 M Imgs. Style GAN2 has obtained the previously lowest FID value of 3.39 after 88 M Imgs, 80 times as many as needed by Projected GAN. Similar speed-ups are also realized for all other large datasets as shown in Table 3. Interestingly, when training longer on FFHQ (39 M Imgs), we observe further improvements of Projected GAN to an FID of 2.2. Note that all five datasets represent very different objects in various scenes. This demonstrates that the performance gain is robust to the choice of the dataset, although the feature network is trained only on Image Net. It is important to note that the main improvements are based on improved sample diversity as indicated by recall which we report in the appendix. The improvement in diversity is most notable on large datasets, e.g., LSUN church, where the image fidelity appears to be similar to Style GAN. 5.3 Small Datasets To further evaluate our method in the few-shot setting, we compare against Style GAN2-ADA and Fast GAN on art paintings from Wiki Art (1000 images; wikiart.org), Oxford Flowers (1360 images) [42], photographs of landscapes (4319 images; flickr.com), Animal Face-Dog (389 images) [57] and Pokemon (833 images; pokemon.com). Further, we report results on high-resolution versions of Pokemon and Art-Painting (10242). Lastly, we evaluate on AFHQ-Cat, -Dog and -Wild at 5122 [5]. The AFHQ datasets contain 5k closeups per category cat, dog, or wildlife. We do not have a license to re-distribute these datasets, but we provide the URLs to enable reproducibility, similar to [35]. Projected GAN outperforms all baselines in terms of FID values by a significant margin on all datasets and all resolutions as shown in Table 3. Remarkably, our model beats the prior state-of-the-art on all datasets (2562) after observing fewer than 600k images. For Animal Face-Dog, the Projected GAN surpasses the previously best FID after only 20k images. One might argue that the Efficient Net used as feature network facilitates data generation for the animal datasets as Efficient Net is trained on Image Net which contains many animal classes (e.g., 120 classes for dog breeds). However, it is interesting to observe that Projected GANs also achieve state-of-the-art FID on Pokemon and Art Painting though these datasets differ significantly from Image Net. This evidences the generality of Image Net features. For the high-resolution datasets, Projected GANs achieve the same FID value many times faster than the best baselines, e.g., ten times faster than Style GAN2-ADA on AFHQCat or four times faster than Fast GAN on Pokemon. We remark that F and Dl generalize to any resolution as they are fully convolutional. Figure 6: Real samples (top rows) vs. samples by Projected GAN (bottom rows). Datasets (top left to bottom right): CLEVR (2562), LSUN church (2562), Art Painting (2562), Landscapes (2562), AFHQ-wild (5122), Pokemon (2562), AFHQ-dog (5122), AFHQ-cat (5122). FID Imgs FID Imgs FID Imgs FID Imgs FID Imgs Large Datasets (2562) CLEVR FFHQ Cityscapes Bedroom Church SAGAN [69] 26.04 10 M 16.21 10 M 12.81 10 M 14.06 10 M 6.15 10 M STYLEGAN2-ADA [25] 10.17 10 M 7.32 10 M 8.35 10 M 11.53 10 M 5.85 10 M GANSFORMERS [20] 9.24 10 M 7.42 10 M 5.23 10 M 6.15 10 M 5.47 10 M FASTGAN [35] 3.24 10 M 12.69 10 M 8.78 1.8 M 8.24 4.8 M 8.43 8.9 M PROJECTED GAN 0.89 4.5 M 3.39 7.1 M 3.41 1.7 M 1.52 5.2 M 1.59 9.2 M PROJECTED GAN* 3.39 0.5 M 3.56 7.0 M 4.60 1.1 M 2.58 1.5 M 3.18 1.1 M STYLEGAN2* [25,26,68] 5.05 25 M 3.62 25 M - - 2.65 70 M 3.39 88 M Small Datasets (2562) Art Painting Landscape Animal Face Flowers Pokemon STYLEGAN2-ADA [25] 43.07 3.2 M 15.99 6.3 M 60.90 2.2 M 21.66 3.8 M 40.38 3.4 M FASTGAN [35] 44.02 0.7 M 16.44 1.8 M 62.11 0.2 M 26.23 0.8 M 81.86 2.5 M PROJECTED GAN 27.96 0.8 M 6.92 3.5 M 17.88 10 M 13.86 1.8 M 26.36 0.8 M PROJECTED GAN* 40.22 0.2 M 14.99 0.6 M 58.07 0.02 M 21.60 0.2 M 36.57 0.3 M Art Painting Pokemon AFHQ-Cat AFHQ-Dog AFHQ-Wild STYLEGAN2-ADA [25] 41.69 1.0 M 56.76 0.6 M 3.55 10 M 7.40 10 M 3.05 10 M FASTGAN [35] 46.71 0.8 M 56.46 0.8 M 4.69 1.1 M 13.09 1.6 M 3.14 1.6 M PROJECTED GAN 32.07 0.9 M 33.96 1.3 M 2.16 3.7 M 4.52 3.8 M 2.17 5.4 M PROJECTED GAN* 40.33 0.2 M 53.74 0.2 M 3.53 1.0 M 7.10 0.9 M 3.03 1.6 M Table 3: Quantitative Results. Projected GAN* reports the point where our approach surpasses the state-of-the-art. Style GAN2* obtains the lowest FID in previous literature if trained long enough. 6 Discussion and Future Work Figure 7: "Floating Heads" Figure 8: Artifacts on FFHQ. While we achieve low FID on all datasets, we also identify two systematic failure cases: As depicted in Fig. 7, we sometimes observe floating heads on AFHQ. In a few samples, the animals appear in high quality but resemble cutouts on blurry or bland backgrounds. We hypothesize that generating a realistic background and image composition is less critical when a prominent object is already depicted. This hypothesis follows from the fact that we used image classification models for the projection, which have been shown to only marginally reduce in accuracy when applied on images of objects with removed background [66]. On FFHQ, projected GAN sometimes produces poor-quality samples with wrong proportions and artifacts, even at state-of-the-art FID, see Fig. 8. In terms of generators, Style GAN is more challenging to tune and does not profit as much from projected training. The Fast GAN generator is fast to optimize but simultaneously produced unrealistic samples in some parts of the latent space a problem that could be solved by a mapping network similar to Style GAN. Hence, we speculate that unifying the strengths of both architectures in combination with projected training might improve performance further. Moreover, our study of different pretrained networks indicates that efficient models are especially suitable for projected GAN training. Exploring this connection in-depth, and in general, determining desirable feature space properties opens up exciting new research opportunities. Lastly, our work advances efficiency for generative models. More efficient models lower the barrier of computational effort needed for generating realistic images. A lower barrier facilitates malignant use of generative models (e.g., deep fakes ) while simultaneously also democratizing research in this area. Acknowledgments and Disclosure of Funding We acknowledge the financial support by the BMWi in the project KI Delta Learning (project number 19A19013O). Andreas Geiger was supported by the ERC Starting Grant LEGO-3D (850533). Kashyap Chitta was supported by the German Federal Ministry of Education and Research (BMBF): Tübingen AI Center, FKZ: 01IS18039B and the International Max Planck Research School for Intelligent Systems (IMPRS-IS). Jens Müller received funding by the Heidelberg Collaboratory for Image Processing (HCI). We thank the Center for Information Services and High Performance Computing (ZIH) at Dresden University of Technology for generous allocations of computation time. Lastly, we would like to thank Vanessa Sauer for her general support. [1] I. Albuquerque, J. Monteiro, T. Doan, B. Considine, T. Falk, and I. Mitliagkas. Multi-objective training of generative adversarial networks with multiple discriminators. In Proc. of the International Conf. on Machine learning (ICML), 2019. 2, 3 [2] M. Arjovsky and L. Bottou. Towards principled methods for training generative adversarial networks. In Proc. of the International Conf. on Learning Representations (ICLR), 2017. 2 [3] M. Binkowski, D. J. Sutherland, M. Arbel, and A. Gretton. Demystifying MMD gans. In Proc. of the International Conf. on Learning Representations (ICLR), 2018. 6 [4] A. Brock, J. Donahue, and K. Simonyan. Large scale GAN training for high fidelity natural image synthesis. In Proc. of the International Conf. on Learning Representations (ICLR), 2019. 1, 2 [5] Y. Choi, Y. Uh, J. Yoo, and J.-W. Ha. Stargan v2: Diverse image synthesis for multiple domains. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 1, 8 [6] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele. The cityscapes dataset for semantic urban scene understanding. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2016. 8 [7] J. Deng, W. Dong, R. Socher, L. jia Li, K. Li, and L. Fei-fei. Imagenet: A large-scale hierarchical image database. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2009. 4 [8] T. Doan, J. Monteiro, I. Albuquerque, B. Mazoure, A. Durand, J. Pineau, and R. D. Hjelm. Online adaptative curriculum learning for gans. In Proc. of the Conf. on Artificial Intelligence (AAAI), 2019. 2 [9] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In Proc. of the International Conf. on Learning Representations (ICLR), 2021. 4 [10] A. Dosovitskiy and T. Brox. Generating images with perceptual similarity metrics based on deep networks. In Advances in Neural Information Processing Systems (NIPS), 2016. 2 [11] A. Dosovitskiy and T. Brox. Inverting visual representations with convolutional networks. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2016. 5 [12] D. Dowson and B. Landau. The fréchet distance between multivariate normal distributions. Journal of multivariate analysis, 1982. 5 [13] I. Durugkar, I. Gemp, and S. Mahadevan. Generative multi-adversarial networks. In Proc. of the International Conf. on Learning Representations (ICLR), 2017. 2, 3 [14] P. Esser, R. Rombach, and B. Ommer. Taming transformers for high-resolution image synthesis. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 2 [15] H. Ham, T. J. Jun, and D. Kim. Unbalanced gans: Pre-training the generator of generative adversarial network using variational autoencoder. ar Xiv.org, 2002.02112, 2020. 2 [16] K. He, X. Zhang, S. Ren, and J. Sun. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proc. of the IEEE International Conf. on Computer Vision (ICCV), 2015. 4 [17] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems (NIPS), 2017. 4, 5, 6 [18] J. Howard and S. Ruder. Universal language model fine-tuning for text classification. In Association for Computational Linguistics (ACL), 2018. 2 [19] X. Huang, Y. Li, O. Poursaeed, J. Hopcroft, and S. Belongie. Stacked generative adversarial networks. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2017. 2 [20] D. A. Hudson and C. L. Zitnick. Generative adversarial transformers. ar Xiv.org, 2103.01209, 2021. 6, 7, 8, 9 [21] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses for real-time style transfer and superresolution. In Proc. of the European Conf. on Computer Vision (ECCV), 2016. 2 [22] J. Johnson, B. Hariharan, L. van der Maaten, L. Fei-Fei, C. Lawrence Zitnick, and R. Girshick. Clevr: A diagnostic dataset for compositional language and elementary visual reasoning. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2017. 7 [23] A. Karnewar and O. Wang. Msg-gan: Multi-scale gradients for generative adversarial networks. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 2 [24] T. Karras, T. Aila, S. Laine, and J. Lehtinen. Progressive growing of GANs for improved quality, stability, and variation. In Proc. of the International Conf. on Learning Representations (ICLR), 2018. 2 [25] T. Karras, M. Aittala, J. Hellsten, S. Laine, J. Lehtinen, and T. Aila. Training generative adversarial networks with limited data. In Advances in Neural Information Processing Systems (Neur IPS), 2020. 2, 3, 4, 6, 7, 9 [26] T. Karras, S. Laine, and T. Aila. A style-based generator architecture for generative adversarial networks. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019. 1, 2, 6, 8, 9 [27] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, and T. Aila. Analyzing and improving the image quality of Style GAN. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 2, 6 [28] D. P. Kingma and P. Dhariwal. Glow: Generative flow with invertible 1x1 convolutions. In Advances in Neural Information Processing Systems (Neur IPS), 2018. 4 [29] A. Kolesnikov, L. Beyer, X. Zhai, J. Puigcerver, J. Yung, S. Gelly, and N. Houlsby. Big transfer (bit): General visual representation learning. In Proc. of the European Conf. on Computer Vision (ECCV), 2020. 1 [30] S. Kornblith, J. Shlens, and Q. V. Le. Do better imagenet models transfer better? In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019. 1 [31] K. Kurach, M. Lucic, X. Zhai, M. Michalski, and S. Gelly. The gan landscape: Losses, architectures, regularization, and normalization. In Proc. of the International Conf. on Machine learning (ICML) Workshops, 2018. 1 [32] C. Ledig, L. Theis, F. Huszar, J. Caballero, A. Cunningham, A. Acosta, A. P. Aitken, A. Tejani, J. Totz, Z. Wang, and W. Shi. Photo-realistic single image super-resolution using a generative adversarial network. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2017. 2 [33] J. H. Lim and J. C. Ye. Geometric gan. ar Xiv.org, 1705.02894, 2017. 4 [34] T. Lin, P. Dollár, R. B. Girshick, K. He, B. Hariharan, and S. J. Belongie. Feature pyramid networks for object detection. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2017. 2 [35] B. Liu, Y. Zhu, K. Song, and A. Elgammal. Towards faster and stabilized gan training for highfidelity few-shot image synthesis. In Proc. of the International Conf. on Learning Representations (ICLR), 2021. 4, 6, 8, 9 [36] L. Mescheder, A. Geiger, and S. Nowozin. Which training methods for gans do actually converge? In Proc. of the International Conf. on Machine learning (ICML), 2018. 1 [37] T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida. Spectral normalization for generative adversarial networks. In Proc. of the International Conf. on Learning Representations (ICLR), 2018. 3 [38] S. Mo, M. Cho, and J. Shin. Freeze the discriminator: a simple baseline for fine-tuning gans. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) Workshops, 2020. 2 [39] S. Morozov, A. Voynov, and A. Babenko. On self-supervised image representations for GAN evaluation. In Proc. of the International Conf. on Learning Representations (ICLR), 2021. 6 [40] B. Neyshabur, S. Bhojanapalli, and A. Chakrabarti. Stabilizing gan training with multiple random projections. ar Xiv.org, 1705.0783, 2017. 2, 3 [41] A. Nguyen, J. Yosinski, Y. Bengio, A. Dosovitskiy, and J. Clune. Plug & play generative networks: Conditional iterative generation of images in latent space. ar Xiv.org, 1612.00005, 2016. 2 [42] M.-E. Nilsback and A. Zisserman. Automated flower classification over a large number of classes. In Proc. Indian Conf. on Computer Vision, Graphics & Image Processing, 2008. 8 [43] T. Park, M.-Y. Liu, T.-C. Wang, and J.-Y. Zhu. Semantic image synthesis with spatially-adaptive normalization. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019. 2 [44] O. Patashnik, Z. Wu, E. Shechtman, D. Cohen-Or, and D. Lischinski. Styleclip: Text-driven manipulation of stylegan imagery. ar Xiv.org, 2103.17249, 2021. 2 [45] M. E. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, and L. Zettlemoyer. Deep contextualized word representations. In Conf. of the North American Chapter of the Association for Computational Linguistics (NAACL), 2018. 2 [46] A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al. Learning transferable visual models from natural language supervision. ar Xiv.org, 2103.00020, 2021. 4 [47] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever. Improving language understanding by generative pre-training, 2018. 2 [48] R. Ranftl, K. Lasinger, D. Hafner, K. Schindler, and V. Koltun. Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. on Pattern Analysis and Machine Intelligence (PAMI), 2020. 1 [49] S. R. Richter, H. A. Al Haija, and V. Koltun. Enhancing photorealism enhancement. ar Xiv.org, 2105.04619, 2021. 2 [50] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2015. 4 [51] M. S. Sajjadi, O. Bachem, M. Lucic, O. Bousquet, and S. Gelly. Assessing generative models via precision and recall. In Advances in Neural Information Processing Systems (Neur IPS), 2018. 6 [52] M. S. Sajjadi, B. Scholkopf, and M. Hirsch. Enhancenet: Single image super-resolution through automated texture synthesis. In Proc. of the IEEE International Conf. on Computer Vision (ICCV), 2017. 2 [53] C. N. d. Santos, Y. Mroueh, I. Padhi, and P. Dognin. Learning implicit generative models by matching perceptual features. In Proc. of the IEEE International Conf. on Computer Vision (ICCV), 2019. 2 [54] A. Sauer and A. Geiger. Counterfactual generative networks. In Proc. of the International Conf. on Learning Representations (ICLR), 2021. 2 [55] E. Schonfeld, B. Schiele, and A. Khoreva. A u-net based discriminator for generative adversarial networks. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 2 [56] A. Shocher, Y. Gandelsman, I. Mosseri, M. Yarom, M. Irani, W. T. Freeman, and T. Dekel. Semantic pyramid for image generation. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020. 2 [57] Z. Si and S.-C. Zhu. Learning hybrid image templates (hit) by information projection. IEEE Trans. on Pattern Analysis and Machine Intelligence (PAMI), 2011. 8 [58] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In Proc. of the International Conf. on Learning Representations (ICLR), 2015. 2 [59] D. Sungatullina, E. Zakharov, D. Ulyanov, and V. Lempitsky. Image manipulation with perceptual discriminators. In Proc. of the European Conf. on Computer Vision (ECCV), 2018. 2, 5 [60] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. E. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich. Going deeper with convolutions. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2015. 4 [61] M. Tan and Q. Le. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proc. of the International Conf. on Machine learning (ICML), 2019. 4 [62] H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. Jégou. Training dataefficient image transformers & distillation through attention. ar Xiv.org, 2012.12877, 2020. 4 [63] N.-T. Tran, V.-H. Tran, N.-B. Nguyen, T.-K. Nguyen, and N.-M. Cheung. On data augmentation for gan training. IEEE Trans. on Image Processing (TIP), 2021. 2, 4 [64] T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, and B. Catanzaro. High-resolution image synthesis and semantic manipulation with conditional gans. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2018. 2 [65] Y. Wang, C. Wu, L. Herranz, J. van de Weijer, A. Gonzalez-Garcia, and B. Raducanu. Transferring gans: generating images from limited data. In Proc. of the European Conf. on Computer Vision (ECCV), 2018. 2 [66] K. Xiao, L. Engstrom, A. Ilyas, and A. Madry. Noise or signal: The role of image backgrounds in object recognition. In Proc. of the International Conf. on Learning Representations (ICLR), 2021. 10 [67] F. Yu, A. Seff, Y. Zhang, S. Song, T. Funkhouser, and J. Xiao. Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. ar Xiv.org, 1506.03365, 2015. 4, 8 [68] N. Yu, G. Liu, A. Dundar, A. Tao, B. Catanzaro, L. Davis, and M. Fritz. Dual contrastive loss and attention for gans. ar Xiv.org, 2103.16748, 2021. 7, 9 [69] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena. Self-attention generative adversarial networks. In Proc. of the International Conf. on Machine learning (ICML), 2019. 2, 8, 9 [70] J. Zhao, M. Mathieu, and Y. Le Cun. Energy-based generative adversarial network. In Proc. of the International Conf. on Learning Representations (ICLR), 2017. 2 [71] M. Zhao, Y. Cong, and L. Carin. On leveraging pretrained gans for generation with limited data. In Proc. of the International Conf. on Machine learning (ICML), 2020. 2 [72] S. Zhao, Z. Liu, J. Lin, J.-Y. Zhu, and S. Han. Differentiable augmentation for data-efficient gan training. In Advances in Neural Information Processing Systems (Neur IPS), 2020. 2, 4, 7 [73] Z. Zhao, Z. Zhang, T. Chen, S. Singh, and H. Zhang. Image augmentations for gan training. ar Xiv.org, 2006.02595, 2020. 2, 4 [74] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycleconsistent adversarial networks. In Proc. of the IEEE International Conf. on Computer Vision (ICCV), 2017. 2