# discriminator_optimal_transport__c87258c4.pdf Discriminator optimal transport Akinori Tanaka Mathematical Science Team, RIKEN Center for Advanced Intelligence Project (AIP) 1-4-1 Nihonbashi, Chuo-ku, Tokyo 103-0027, Japan Interdisciplinary Theoretical and Mathematical Sciences Program (i THEMS), RIKEN 2-1 Hirosawa, Wako, Saitama 351-0198, Japan Department of Mathematics, Faculty of Science and Technology, Keio University 3-14-1 Hiyoshi, Kouhoku-ku, Yokohama 223-8522, Japan akinori.tanaka@riken.jp Within a broad class of generative adversarial networks, we show that discriminator optimization process increases a lower bound of the dual cost function for the Wasserstein distance between the target distribution p and the generator distribution p G. It implies that the trained discriminator can approximate optimal transport (OT) from p G to p. Based on some experiments and a bit of OT theory, we propose discriminator optimal transport (DOT) scheme to improve generated images. We show that it improves inception score and FID calculated by unconditional GAN trained by CIFAR-10, STL-10 and a public pre-trained model of conditional GAN trained by Image Net. 1 Introduction Generative Adversarial Network (GAN) [1] is one of recent promising generative models. In this context, we prepare two networks, a generator G and a discriminator D. G generates fake samples G(z) from noise z and tries to fool D. D classifies real sample x and fake samples y = G(z). In the training phase, we update them alternatingly until it reaches to an equilibrium state. In general, however, the training process is unstable and requires tuning of hyperparameters. Since from the first successful implementation by convolutional neural nets [2], most literatures concentrate on how to improve the unstable optimization procedures including changing objective functions [3, 4, 5, 6, 7, 8], adding penalty terms [9, 10, 11], techniques on optimization precesses themselves [12, 13, 14, 15], inserting new layers to the network [16, 17], and others we cannot list here completely. Even if one can make the optimization relatively stable and succeed in getting G around an equilibrium, it sometimes fails to generate meaningful images. Bad images may include some unwanted structures like unnecessary shadows, strange symbols, and blurred edges of objects. For example, see generated images surrounded by blue lines in Figure 1. These problems may be fixed by scaling up the network structure and the optimization process. Generically speaking, however, it needs large scale computational resources, and if one wants to apply GAN to individual tasks by making use of more compact devices, the above problem looks inevitable and crucial. There is another problem. In many cases, we discard the trained discriminator D after the training. This situation is in contrast to other latent space generative models. For example, variational autoencoder (VAE) [18] is also composed of two distinct networks, an encoder network and a decoder network. We can utilize both of them after the training: the encoder can be used as a data compressor, and the decoder can be regarded as a generator. Compared to this situation, it sounds wasteful to use only G after the GAN training. 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. Figure 1: Each left image (blue): a sample from generator G. Each right image (red): the sample modified by our algorithm based on discriminator D. We use here the trained model available on https://github.com/pfnet-research/sngan_projection . From this viewpoint, it would be natural to ask how to use trained models G and D efficiently. Recent related works in the same spirit are discriminator rejection sampling (DRS) [19] and Metropolis-Hastings GAN (MH-GAN) [20]. In each case, they use the generator-induced distribution p G as a proposal distribution, and approximate acceptance ratio of the proposed sample based on the trained D. Intuitively, generated image y = G(z) is accepted if the value D(y) is relatively large, otherwise it is rejected. They show its theoretical backgrounds, and it actually improve scores on generated images in practice. In this paper, we try a similar but more active approach, i.e. improving generated image y = G(z) directly by adding δy to y such that D(y +δy) > D(y). The optimal transport (OT) theory guarantees that the improved samples can be regarded as approximate samples from the target distribution p. More concretely, our contributions are Proposal of the discriminator optimal transport (DOT) based on the fact that the objective function for D provides lower bound of the dual cost function for the Wasserstein distance between p and p G. Suggesting approximated algorithms and verifying that they improve Earth Mover s distance (EMD) [21], inception score [13] and Fréchet inception distance (FID) [15]. Pointing out a generality on DOT, i.e. if the G s output domain is same as the D s input domain, then we can use any pair of trained G and D to improve generated samples. In addition, we show some results on experiment comparing DOT and a naive method of improving sample just by increasing the value of D, under a fair setting. One can download our codes from https://github.com/Akinori Tanaka-phys/DOT. 2 Background 2.1 Generative Adversarial Nets Throughout this paper, we regard an image sample as a vector in certain Euclidean space: x X. We name latent space as Z and a prior distribution on it as p Z(z). The ultimate goal of the GAN is making generator G : Z X whose push-foward of the prior p G(x) = Z p Z(z)δ ( x G(z) ) dz reproduces data-generating probability density p(x). To achieve it, discriminator D : X R and objective functions, VD(G, D) = Ex p [f( D(x))] + Ey p G [f(D(y))] , (1) VG(G, D) = Ey p G [g(D(y))] , (2) are introduced. Choice of functions f and g corresponds to choice of GAN update algorithm as explained below. Practically, G and D are parametric models Gθ and Dφ, and they are alternatingly updated as φ φ + ϵ φVD(Gθ, Dφ), (3) θ θ ϵ θVG(Gθ, Dφ), (4) until the updating dynamics reaches to an equilibrium. One of well know choices for f and g is f(u) = log(1 + eu) g(u) = f( u). (5) Theoretically speaking, it seems better to take g(u) = f(u), which is called minimax GAN [22] to guarantee p G = p at the equilibrium as proved in [1]. However, it is well known that taking (5), called non-saturating GAN, enjoys better performance practically. As an alternative, we can choose the following f and g [6, 4]: f(u) = max(0, 1 u), g(u) = u. (6) It is also known to be relatively stable. In addition to it, p G = p at an equilibrium is proved at least in the theoretically ideal situation. Another famous choice is taking f(u) = u, g(u) = u. (7) The resultant GAN is called WGAN [5]. We use (7) with gradient penalty (WGAN-GP) [9] in our experiment. WGAN is related to the concept of the optimal transport (OT) which we review below, so one might think our method is available only when we use WGAN. But we would like to emphasize that such OT approach is also useful even when we take GANs described by (5) and (6) as we will show later. 2.2 Spectral normalization Spectral normalization (SN) [16] is one of standard normalizations on neural network weights to stabilize training process of GANs. To explain it, let us define a norm for function called Lipschitz norm, ||f||Lip := sup {||f(x) f(y)||2 x = y } . (8) For example, ||Re LU||Lip = ||l Re LU||Lip = 1 because their maximum gradient is 1. For linear transformation l W,b with weight matrix W and bias b, the norm ||l W,b||Lip is equal to the maximum singular value σ(W). Spectral normalization on l W,b is defined by dividing the weight W in the linear transform by the σ(W): SN(l W,b) = l W/σ(W ),b. (9) By definition, it enjoys ||l W/σ(W )||Lip = 1. If we focus on neural networks, estimation of the upper bound of the norm is relatively easy because of the following property1: ||f g||Lip ||f||Lip ||g||Lip. (10) For example, suppose fnn is a neural network with Re LU or l Re LU activations and spectral normalizations: fnn(x) = SN l WL f SN l WL 1 SN l W1(x), then the Lipschitz norm is bounded by one: l=1 ||l Wl/σ(Wl)||Lip = 1 (11) Thanks to this Lipschitz nature, the normalized network gradient behaves mild during repeating updates (3) and (4), and as a result, it stabilizes the wild and dynamic optimization process of GANs. 1 This inequality can be understood as follows. Naively, the norm (8) is defined by the maximum gradient between two different points. Suppose x1 and x2 realizing maximum of gradient for g and u1 and u2 are points for f, then the (RHS) of the inequality (10) is equal to ||f(u1) f(u2)||2/||u1 u2||2 ||g(x1) g(x2)||2/||x1 x2||2. If g(xi) = ui, it reduces to the (LHS) of the (10), but the condition is not satisfied in general, and the (RHS) takes a larger value than (LHS). This observation is actually important to the later part of this paper because estimation of the norm based on the inequality seems to be overestimated in many cases. 2.3 Optimal transport Another important background in this paper is optimal transport. Suppose there are two probability densities, p(x) and q(y) where x, y X. Let us consider the cost for transporting one unit of mass from x p to y q. The optimal cost is called Wasserstein distance. Throughout this paper, we focus on the Wasserstein distance defined by l2-norm cost ||x y||2: W(p, q) = min π Π(p,q) ( E(x,y) π [ ||x y||2 ]) . (12) π means joint probability for transportation between x and y. To realize it, we need to restrict π satisfying marginality conditions, dx π(x, y) = q(y), dy π(x, y) = p(x). (13) An optimal π satisfies W(p, q) = E(x,y) π [||x y||2], and it realizes the most effective transport between two probability densities under the l2 cost. Interestingly, (12) has the dual form W(p, q) = max || D||Lip 1 ( Ex p [ D(x) ] Ey q [ D(y) ]) . (14) The duality is called Kantorovich-Rubinstein duality [23, 24]. Note that ||f||Lip is defined in (8), and the dual variable D should satisfy Lipschitz continuity condition || D||Lip 1. One may wonder whether any relationship between the optimal transport plan π and the optimal dual variable D exists or not. The following theorem is an answer and it plays a key role in this paper. Theorem 1 Suppose π and D are optimal solutions of the primal (12) and the dual (14) problem, respectively. If π is deterministic optimal transport described by a certain automorphism2 T : X X, then the following equations are satisfied: ||D ||Lip = 1, (15) T(y) = arg min x { ||x y||2 D (x) } , (16) p(x) = dy δ ( x T(y) ) q(y). (17) (Proof) It can be proved by combining well know facts. See Supplementary Materials. 3 Discriminator optimal transport If we apply the spectral normalization on a discriminator D, it satisfies ||D||Lip = K with a certain real number K. By redefining it to D = D/K, it becomes 1-Lipschitz || D||Lip = 1. It reminds us the equation (15), and one may expect a connection between OT and GAN. In fact, we can show the following theorem: Theorem 2 Each objective function of GAN using logistic (5), or hinge (6), or identity (7) loss with gradient penalty, provides lower bound of the mean discrepancy of D = D/K between p and p G: VD(G, D) K ( Ex p [ D(x) ] Ey p G [ D(y) ]) . (18) (Proof) See Supplementary Materials. In practical optimization process of GAN, K could change its value during the training process, but it stays almost constant at least approximately as explained below. 2 It is equivalent to assume there exists an unique solution of the corresponding Monge problem: ( Ey q [ ||T(y) y||2 ]) , constrained by (17). Reconstructing T from π without any assumption is a subtle problem and only guaranteed within strictly convex cost functions [25]. Unfortunately, it is not satisfied in our l2 cost. However, there is a known method [26] to find a solution based on relaxing the cost to strictly convex cost ||x y||1+ϵ 2 with ϵ > 0. In our experiments, DOT works only when ||x y||2 is small enough for given y. In this case, there is no big difference between ||x y||2 and ||x y||1+ϵ 2 , and it suggests DOT approximates their solution. Figure 2: Logs of inception score (left), approximated Lipschitz constant of D (middle), and approximated Lipschitz constant of D G (right) on each GAN trained with CIFAR-10. Approximated Lipschitz constants are calculated by random 500 pair samples. Errorbars are plotted within 1σ by 500 trials. 3.1 Discriminator Optimal Transport (ideal version) The inequality (18) implies that the update (3) of D during GANs training maximizes the lower bound of the objective in (14), the dual form of the Wasserstein distance. In this sense, the optimization of D in (3) can be regarded solving the problem (14) approximately3. If we apply (16) with D D = D/K, the following transport of given y p G TD(y) = arg min x { ||x y||2 1 is expected to recover the sampling from the target distribution p thanks to the equality (17). 3.2 Discriminator Optimal Transport (practical version) To check whether K changes drastically or not during the GAN updates, we calculate approximated Lipschitz constants defined by Keff = max {|D(x) D(y)| x, y p G } , (20) keff = max {|D G(z) D G(zy)| z, zy p Z } , (21) in each 5,000 iteration on GAN training with CIFAR-10 data with DCGAN models explained in Supplementary Materials. As plotted in Figure 2, both of them do not increase drastically. It is worth to mention that the naive upper bound of the Lipschitz constant like (11) turn to be overestimated. For example, SNGAN has the naive upper bound 1, but (20) stays around 0.08 in Figure 2. Target space DOT Based on these facts, we conclude that trained discriminators can approximate the optimal transport (16) by T eff D (y) = arg min x { ||x y||2 1 Keff D(x) } . (22) As a preliminary experiment, we apply DOT to WGAN-GP trained by 25gaussians dataset and swissroll dataset. We use the gradient descent method shown in Algorithm 1 to search transported point T eff D (y) for given y p G. In Figure 3, we compare the DOT samples and naively transported samples by the discriminator which is implemented by replacing the gradient in Algorithm 1 to 1 Keff x D(x) , i.e. just searching x with large D(x) from initial condition x y where y p G. DOT outperforms the naive method qualitatively and quantitatively. On the 25gaussians, one might think 4th naively improved samples are better than 3rd DOT samples. However, the 4th samples are too concentrated and lack the variance around each peak. In fact, the value of the Earth Mover s distance, EMD, which measures how long it is separated from the real samples, shows relatively large value. On the swissroll, 4th samples based on naive transport lack many relevant points close to the original data, and it is trivially bad. On the other hand, one can see that the 3rd DOT samples keep swissroll shape and clean the blurred shape in the original samples by generator. 3 This situation is similar to guarantee VAE [18] objective function which is a lower bound of the likelihood called evidence lower bound (ELBO). Algorithm 1 Target space optimal transport by gradient descent Require: trained D, approximated Keff by (20), sample y, learning rate ϵ and small vector δ Initialize x y for ntrial in range(Nupdates) do x x ϵ x { ||x y + δ||2 1 Keff D(x) } ( δ is for preventing overflow. ) end for return x EMD 0.052(08) 0.052(10) 0.065(11) EMD 0.021(05) 0.020(06) 0.160(22) Figure 3: 2d experiments by using trained model of WGAN-GP. 1,000 samples of, 1st: training samples, 2nd: generated samples by G, 3rd: samples by target space DOT, 4th: samples by naive transport, are plotted. Each EMD value is calculated by 100 trials. The error corresponds to 1σ. We use δ = 0.001. See the supplementary material for more details on this experiment. Latent space DOT The target space DOT works in low dimensional data, but it turns out to be useless once we apply it to higher dimensional data. See Figure 4 for example. Alternative, and more workable idea is regarding D G : Z R as the dual variable for defining Wasserstein distance between pullback of p by G and prior p Z. Latent space OT itself is not a novel idea [27, 28], but there seems to be no literature using trained G and D, to the best of our knowledge. The approximated Lipschitz constants of G D also stay constant as shown in the right sub-figure in Figure 2, so we conclude that T eff D G(zy) = arg min z { ||z zy||2 1 keff D G(z) } approximates optimal transport in latent space. Note that if the prior p Z has non-trivial support, we need to restrict z onto the support during the DOT process. In our algorithm 2, we apply projection of the gradient. One of the major practical priors is normal distribution N(0, ID D) where D is the latent space dimension. If D is large, it is well known that the support is concentrated on (D 1)-dimensional sphere with radius D, so the projection of the gradient g can be calculated by g (g z)z/ D approximately. Even if we skip this procedure, transported images may look improved, but it downgrades inception scores and FIDs. Algorithm 2 Latent space optimal transport by gradient descent Require: trained G and D, approximated keff, sample zy, learning rate ϵ, and small vector δ Initialize z zy for ntrial in range(Nupdates) do g = z { ||z zy + δ||2 1 keff D G(z) } ( δ is for preventing overflow. ) if noise is generated by N(0, ID D) then g g (g z)z/ D end if z z ϵg if noise is generated by U([ 1, 1]) then clip z [ 1, 1] end if end for return x = G(z) Figure 4: Target space DOT sample (each left) and latent space DOT sample (each right). The former looks like giving meaningless noises like perturbations in adversarial examples [29]. On the other hand, the latent space DOT samples keep the shape of image, and clean it. 4 Experiments on latent space DOT 4.1 CIFAR-10 and SLT-10 We prepare pre-trained DCGAN models and Res Net models on various settings, and apply latent space DOT. In each case, inception score and FID are improved (Table 1). We can use arbitrary discriminator D to improve scores by fixed G as shown in Table 2. As one can see, DOT really works. But it needs tuning of hyperparameters. First, it is recommended to use small ϵ as possible. A large ϵ may accelerate upgrading, but easily downgrade unless appropriate Nupdates is chosen. Second, we recommend to use keff calculated by using enough number of samples. If not, it becomes relatively small and it also possibly downgrade images. As a shortcut, keff = 1 also works. See Supplementary Materials for details and additional results including comparison to other methods. CIFAR-10 STL-10 bare DOT bare DOT DCGAN WGAN-GP 6.53(08), 27.84 7.45(05), 24.14 8.69(07), 49.94 9.31(07), 44.45 SNGAN(ns) 7.45(09), 20.74 7.97(14), 15.78 8.67(01), 41.18 9.45(13), 34.84 SNGAN(hi) 7.45(08), 20.47 8.02(16), 17.12 8.83(12), 40.10 9.35(12), 34.85 SAGAN(ns) 7.75(07), 25.37 8.50(01), 20.57 8.68(01), 48.23 10.04(14), 41.19 SAGAN(hi) 7.52(06), 25.78 8.38(05), 21.21 9.29(13), 45.79 10.30(21), 40.51 Resnet SAGAN(ns) 7.74(09), 22.13 8.49(13), 20.22 9.33(08), 41.91 10.03(14), 39.48 SAGAN(hi) 7.85(11), 21.53 8.50(12), 19.71 Table 1: (Inception score, FID) by usual sampling (bare) and DOT: Models in [16] and self-attention layer [17] are used. (ns) and (hi) mean models trained by (5) and (6). ϵ = 0.01 SGD is applied 20 times for CIDAR-10 and 10 times for STL-10. keff is calculated by 100 samples and δ = 0.001. D without D WGAN-gp SNGAN(ns) SNGAN(hi) SAGAN(ns) SAGAN(hi) IS 7.52(06) 8.03(11) 8.22(07) 8.38(07) 8.36(12) 8.38(05) FID 25.78 24.47 21.45 23.03 21.07 21.21 Table 2: Results on scores by GSAGAN(ns) after latent space DOT using each D in different training scheme using CIFAR-10 within DCGAN architecture. Parameters for DOT are same in Table 1. Figure 5: Left images surrounded by blue lines are samples from the conditional generator. The number of updates Nupdates for DOT increases along horizontal axis. Right Images surrounded by red lines corresponds after 30 times updates with Adam (α, β1, β2) = (0.01, 0, 0.9) and keff(y) = 1. 4.2 Image Net Conditional version of latent space DOT In this section, we show results on Image Net dataset. As pre-trained models, we utilize a pair of public models (G, D) [30] of conditional GAN [31]4. In conditional GAN, G and D are networks conditioned by label y. Typical objective function VD is therefore represented by average over the label: VD(G, D) = Ey p(y) [ VD ( G( |y), D( |y) )] . (24) But, once y is fixed, G(z|y) and D(x|y) can be regarded as usual networks with input z and x respectively. So, by repeating our argument so far, DOT in conditional GAN can be written by TG D(zy|y) = argminz { ||z zy||2 1 keff(y)D ( G(z|y) y )} . (25) where keff(y) is approximated Lipschitz constant conditioned by y. It is calculated by keff(y) = max {|D ( G(z|y) y ) D ( G(zy|y) y ) | ||z zy||2 z, zy p Z } . (26) Experiments We apply gradient descent updates with with Adam(α, β1, β2) = (0.01, 0, 0.9). We show results on 4 independent trials in Table 3. It is clear that DOT mildly improve each score. Note that we need some tunings on hyperparameters ϵ, Nupdates as we already commented in 4.1. # updates=0 # updates=4 # updates=16 # updates=32 trial1(keff(y) = 1) 36.40(91), 43.34 36.99(75), 43.01 37.25(84), 42.70 37.61(88), 42.35 trial2(keff(y) = 1) 36.68(59), 43.60 36.26(98), 43.09 36.97(63), 42.85 37.02(73), 42.74 trial3 36.64(63), 43.55 36.87(84), 43.11 37.51(01), 42.43 36.88(79), 42.52 trial4 36.23(98), 43.63 36.49(54), 43.25 37.29(86), 42.67 37.29(07), 42.40 Table 3: (Inception score, FID) for each update. Upper 2 cases are executed by keff(y) = 1 without calculating (26). We use 50 samples for each label y to calculate keff(y) in lower 2 trials. δ = 0.001. 4 These are available on https://github.com/pfnet-research/sngan_projection . Algorithm 3 Latent space conditional optimal transport by gradient descent Require: trained G and D, label y, approximated keff(y), sample zy, learning rate ϵ and small vector δ Initialize z zy for ntrial in range(Nupdate) do g = z { ||z zy + δ||2 1 keff(y)D ( G(z|y) y )} ( δ is for preventing overflow. ) if noise is generated by N(0, ID D) then g g (g z)z/ D end if z z ϵg if noise is generated by U([ 1, 1]) then clip z [ 1, 1] end if end for return x = G(z|y) Evaluation To calculate FID, we use available 798,900 image files in ILSVRC2012 dataset. We reshape each image to the size 299 299 3, feed all images to the public inception model to get the mean vector mw and the covariance matrix Cw in 2,048 dimensional feature space explained in Supplementary Materials. 5 Conclusion In this paper, we show the relevance of discriminator optimal transport (DOT) method on various trained GAN models to improve generated samples. Let us conclude with some comments here. First, DOT objective function in (22) reminds us the objective for making adversarial examples [29]. There is known fast algorithm to make adversarial example making use of the piecewise-linear structure of the Re LU neural network [32]. The method would be also useful for accelerating DOT. Second, latent space DOT can be regarded as improving the prior p Z. A similar idea can be found also in [33]. In the usual context of the GAN, we fix the prior, but it may be possible to train the prior itself simultaneously by making use of the DOT techniques. We leave these as future works. Acknowledgments We would like to thank Asuka Takatsu for fruitful discussion and Kenichi Bannai for careful reading this manuscript. This work was supported by computational resources provided by RIKEN AIP deep learning environment (RAIDEN) and RIKEN i THEMS. [1] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, pages 2672 2680, 2014. [2] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, 2016. [3] Sebastian Nowozin, Botond Cseke, and Ryota Tomioka. f-gan: Training generative neural samplers using variational divergence minimization. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, pages 271 279, 2016. [4] Junbo Jake Zhao, Michaël Mathieu, and Yann Le Cun. Energy-based generative adversarial network. Co RR, abs/1609.03126, 2016. [5] Martín Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein GAN. Co RR, abs/1701.07875, 2017. [6] Jae Hyun Lim and Jong Chul Ye. Geometric GAN. Co RR, abs/1705.02894, 2017. [7] Thomas Unterthiner, Bernhard Nessler, Calvin Seward, Günter Klambauer, Martin Heusel, Hubert Ramsauer, and Sepp Hochreiter. Coulomb gans: Provably optimal nash equilibria via potential fields. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [8] Marc G. Bellemare, Ivo Danihelka, Will Dabney, Shakir Mohamed, Balaji Lakshminarayanan, Stephan Hoyer, and Rémi Munos. The cramer distance as a solution to biased wasserstein gradients. Co RR, abs/1705.10743, 2017. [9] Ishaan Gulrajani, Faruk Ahmed, Martín Arjovsky, Vincent Dumoulin, and Aaron C. Courville. Improved training of wasserstein gans. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4-9 December 2017, Long Beach, CA, USA, pages 5767 5777, 2017. [10] Henning Petzka, Asja Fischer, and Denis Lukovnikov. On the regularization of wasserstein gans. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [11] Xiang Wei, Boqing Gong, Zixia Liu, Wei Lu, and Liqiang Wang. Improving the improved training of wasserstein gans: A consistency term and its dual effect. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [12] Luke Metz, Ben Poole, David Pfau, and Jascha Sohl-Dickstein. Unrolled generative adversarial networks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, 2017. [13] Tim Salimans, Ian J. Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved techniques for training gans. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, pages 2226 2234, 2016. [14] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of gans for improved quality, stability, and variation. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [15] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4-9 December 2017, Long Beach, CA, USA, pages 6626 6637, 2017. [16] Takeru Miyato, Toshiki Kataoka, Masanori Koyama, and Yuichi Yoshida. Spectral normalization for generative adversarial networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [17] Han Zhang, Ian J. Goodfellow, Dimitris N. Metaxas, and Augustus Odena. Self-attention generative adversarial networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 7354 7363, 2019. [18] Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. [19] Samaneh Azadi, Catherine Olsson, Trevor Darrell, Ian J. Goodfellow, and Augustus Odena. Discriminator rejection sampling. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. [20] Ryan D. Turner, Jane Hung, Eric Frank, Yunus Saatchi, and Jason Yosinski. Metropolishastings generative adversarial networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 6345 6353, 2019. [21] R emi Flamary and Nicolas Courty. Pot python optimal transport library, 2017. [22] William Fedus, Mihaela Rosca, Balaji Lakshminarayanan, Andrew M. Dai, Shakir Mohamed, and Ian J. Goodfellow. Many paths to equilibrium: Gans do not need to decrease a divergence at every step. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [23] Cédric Villani. Optimal transport: old and new, volume 338. Springer Science & Business Media, 2008. [24] Gabriel Peyré and Marco Cuturi. Computational optimal transport. Foundations and Trends in Machine Learning, 11(5-6):355 607, 2019. [25] Wilfrid Gangbo and Robert J Mc Cann. The geometry of optimal transportation. Acta Mathematica, 177(2):113 161, 1996. [26] Luis Caffarelli, Mikhail Feldman, and Robert Mc Cann. Constructing optimal maps for monge s transport problem as a limit of strictly convex costs. Journal of the American Mathematical Society, 15(1):1 26, 2002. [27] Eirikur Agustsson, Alexander Sage, Radu Timofte, and Luc Van Gool. Optimal transport maps for distribution preserving operations on latent spaces of generative models. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. [28] Tim Salimans, Han Zhang, Alec Radford, and Dimitris N. Metaxas. Improving gans using optimal transport. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [29] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian J. Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. [30] Takeru Miyato and Masanori Koyama. cgans with projection discriminator. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [31] Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. Co RR, abs/1411.1784, 2014. [32] Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. [33] Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale GAN training for high fidelity natural image synthesis. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019.